2/ Horizon World has terrible graphics, but Self-Supervised Learning is foundational AI tech that'll allow multi-modal training beyond labeled data, making speech-to-3D object creation in immersive worlds a reality. Will this yield good world design? TBD.
3/ Project CAIRaoke is a new conversational assistant synthesizing Natural Language Understanding, Dialog State Tracking, Dialog Policy, Management, & Natural Language Generation.
They say it's "deeply contextual & personal."
Video:
4/ This announcement is another incremental step towards @Meta's vision of "contextually-aware AI" that can only be achieved via "egocentric data capture" & "egocentric perception."
Essentially omniscient AI that can respond to your situation & context.
5/ Zuckerberg mentioned #Ego4D, which is 3.7k hrs of egocentric video data in collab w dozen universities to make AI training benchmarks around Episodic Memory, Hand-Object Interactions, AV Diarization, Social, & Forecasting.
6/ I understand the utility of #Ego4D for AR & how it's the next frontier of Computer Vision + AI research, but it's still somewhat creepy as it's also in the context of @meta's omniscient Contextually-Aware AI in the service of surveillance capitalism.
7/ Contextually-Aware AI #Ego4D Benchmarks = Surveillance
Episodic Memory
Querying Memory
Query Construction
Recalling Lives
Hand-Object Interactions
State Changes
AV Diarization
Hearing Words
Tracking Conversations
Social Gaze
Social Dynamics
Forecasting Movements & Interactions
8/ The challenge with Contextually-Aware AI is ensuring it's aligned with Nissenbaum's Contextual Integrity Theory of Privacy ensuring proper flows of information.
How & why is this omniscient-level of data being used & to what ends?
9/ I attended a Meta press briefing on Contextually-Aware AI on 3/18/21, & I paraphrased their goals along with some of my initial reactions to it in this podcast.
SPOILER: I'm not yet convinced of contextually-aware, omnipresent & omniscient AI overlords voicesofvr.com/985-facebook-h…
10/ There's two other projects @Meta announced today. 1. No Language Left Behind: translation system that can learn any language. 2. Universal Speech Translator: real-time speech-to-speech translation across all languages(!!!) which sounds amazing (but never get to 100% accuracy)
11/ Zuckerberg emphasized that some of research that Meta is doing with AI is fundamental as it is being used in other contexts like speaking up MRIs up to 4x:
12/ Zuckerberg claims Meta is commited to building openly & responsibly as AI technologies "deliver the highest levels of privacy & help prevent harm" citing CrypTen.
But not sure how contextually-aware AI will deliver "the highest level of privacy."
It's the opposite of that.
13/ That's the highlights from 15-min intro video Zuckerberg streamed from his page.
There's another 2.5 hour video "Inside the Lab: Building for the Metaverse with AI" on @MetaAI's page digging into more topics.
14/ I watched this 2.5 hour program from Meta on "Building for the Metaverse with AI," and the target audience is potential AI engineers as it was an extended job recruitment pitch, especially at the end (see slides below).
But I'll share some reflections on AI tech & ethics.
15/ Evolution of @Meta's ML architectures is interesting
Past: Supervised Learning & Task-Specific Models
Present: Self-Supervised Learning on Task-Independent Models to Tasks, Product & Tech
Future: SSL with Unified Model & Complex Prediction Tasks
NOTE: Sentiment Analysis & Ads
16/ In Building for the Metaverse, @Meta will combine Self-Supervised Learning, Continual Learning, Reinforcement Learning, & Planning & Reasoning to ALL XR data like biometric sensor data + egocentric data + environmental data to make Contextually-Relevant AI & a "World" Model.
17/ This all encompassing "World" model is so important is because of the types of multi-modal sensor fusion that @Meta will be doing.
For example, at #IEEEVR 2021 @RealityLabs showed how they could extrapolate Eye Gaze from Hand Pose + Head Pose:
18/ Here's a taxonomy of how different XR data & biometric measurements can do psychographic profiling that tracks XR user's actions, mental thoughts, unconscious physiological reactions & emotional states.
What are the mental privacy implications of fusing all of these together?
19/ Human Rights Lawyer @brittanheller gave a great @Gatherverse speech on 5 things XR privacy folks should understand about how XR data is completely different including lack of legal definition of this data.
20/ @Meta first published their Five Pillars of Responsible AI on June 22, 2021 & are diff than their four RI Principles:
Privacy & Security
Fairness & Inclusion
Robustness & Safety
Transparency & Control
Governance & Accountability ai.facebook.com/blog/facebooks…
21/ In April 2021, @Meta published a Privacy Progress Update with 8 Core Principles of how they define privacy, which is all about controlling access to identified data, but there's nothing on physiological XR data or profiling via biometric psychography.
22/ @Meta is still in an old paradigm of defining privacy as identity & control of data rather than contextually-relevant psychographic profiling via biometric data.
They still haven't commented on #NeuroRights like a Right to Mental Privacy.
See my talk
23/ Meta's Five Pillars of Responsible AI is an incremental step to RI best practices, but there's still a long ways to go.
Two talks with more
Towards a Framework for XR Ethics
Sensemaking Frameworks for the Metaverse & XR Ethics
24/ Meta's consistently frames privacy issues back to individual identity & the ways they're protecting personal identifiable information from leaking out in their ML training analysis.
They also oddly often shifted from privacy into open source & transparency tools like CrypTen.
25/ Back to the overall thrust of the presentation, it covered major applications & intentions for @MetaAI's efforts, and how their research is being applied to the Metaverse, which I'll dig into a bit.
They focused primarily on Robotic Embodiment, Creative Apps, & Safety.
26/ Here's a number of graphics showing the faster improving performance of their Self-Supervised Learning approaches vs Supervised Learning with different algorithms progressing over the years.
27/ Here's a number of different haptics & robotics applications of Self-Supervised Learning.
[NOTE: Searching for Meta vs Facebook has killed their SEO in trying to track down the original papers or references of some of this]
28/ Here's a series of graphics that shows what modules within an Classical ConvAI Pipeline that Meta's new CAIRaoke Model replaces.
[NOTE: yes, it is pronounced "karaoke"]
Additionally, the CAIRaoke output is recursively fed back in as an iterative input with sample assertions.
29/ @Meta showed a speculative AR design demonstrating how creepy Contextually-Aware AI might be.
It knows you already added enough salt & automatically buys more salt for you.
Has a memo from your mom + shows how she sliced Habeneros thinly, unclear if AI is extrapolating these.
30/ @lexfridman facilitated a really engaging 40-min discussion with pioneering AI researchers Yoshua Bengio & @MetaAI's @ylecun on the pathway to human-level AI, self-supervised learning, & consciousness (Bengio likes Graziano's Attention Schema Theory). facebook.com/MetaAI/videos/…
31/ There was a whole section on language translation that I found interesting. I know how speech-to-tech is not and may never reach 100% accuracy, & so I really wonder how to quantify the accuracy & communication loss of the vision of these universal translators.
32/ They used to use English as an intermediate language to translate into, but they're moving to a system where they can directly translate any language into another language.
It makes me wonder about gaps in contextual knowledge & how anyone will ever detect mistranslations
33/ Some slides describing how Self-Supervised Learning works by taking away information & then requiring the ML to learning abstract schemas to be used for sensemaking & filling in the gaps. This can be iterated on for SSL to learn how to discern patterns more efficiently.
34/ There was a whole section on Responsible AI approaches & projects that @Meta has been working on.
The process of Tracking Fairness Across Meta in AI is a bit of a more robust & fleshed out responsible innovation framework than Reality Labs has talked about so far.
35/ I don't fully understand how they can model & explain SSL models.
'AI Model Cards are a standardized way to document, track, & monitor individual models...'
'AI System Cards how a group of AI System of AI Models & other non-AI techniques work together to accomplish a task.'
36/ Part of my hesitation in diving into running everything on self-learning AI models is it's very easy to create complex systems that you don't fully understand.
I clicked on links quickly tracking down a story & I was auto-blocked by Meta's AI overlords
37/ There's a lot of ways in which both safety & security requirements for moderation for XR AND the challenges of immersive VR & AR content + egocentric data capture are all catalysts for new AI benchmarks & challenges and even sometimes new algorithmic approaches.
38/ Starting to wrap up this thread.
These types of talks by Meta are quite dense to sift through, but they also usually reveal quite a lot of interesting insights into their deeper strategies & philosophies.
Also VR, AR, & the Metaverse are great provocations for AI innovations.
39/ Ultimately, this was a recruitment pitch.
It's the most optimistic story that @Meta tells itself about who they are, what they do, & why.
Creating AI to benefit people & society.
Building the future responsibly.
Ensuring fairness & robustness.
Collaborating with stakeholders.
40/ I have so many disconnects when I hear Zuckerberg say he want to "deliver the highest levels of privacy" while also aspiring to create contextually-aware AI with #Ego4D that can reconstruct memories, track every conversation, & predict your next move.
2/ Zuckerberg saying we're going to be taking another big step forward with VR after walking through a history of VR tech since the Oculus acquisition.
3/ @boztank talking about some ecosystem stats for VR apps.
Philosophically they want to make ecosystem more open & how some apps have onboarded via the App Lab. Will push towards openness.
Pico was acquired by Bytedance (aka @tiktok_us) & they’re taking on Meta as the best standalone VR competitor to the Quest (but only in Europe & Asia to start).
Thread 🧵👇 to digest the livestream & news
2/ My sources of @PICOXR news are the official livestream👆
@CasandChary have the most comprehensive & digest & hands on:
3/ Here's 1st of 4 trailers @PICOXR showed during their Pico 4 announcement livestream.
Pico 4 has color, mixed reality passthrough while Meta is expected to launch similar features with Meta Quest Pro but at a higher price point.
Pico 4 Pro details to be announced at AWE Lisbon
1/ I recorded an epic recap & reviews of the 30 immersive stories in competition at #VeniceImmersive with @vrgeschichten.
We chat about each experience for ~5 min, & so it's a comprehensive break down of this year's selection. voicesofvr.com/1121-venice-im…
20+ interviews coming soon
2/ I'm about to release other 20 podcasts from #VeniceImmersive, and here's a sneak peak of the 26 episodes, which total just over 24 hours worth of coverage.
3/ Here's the Twitter thread of my live coverage from #VeniceImmersive, which starts with an interview with co-curators Liz Rosenthal and Michel Reilhac. voicesofvr.com/1119-venice-im…
I'll be onsite at #VeniceImmersive covering the fest & this is will be my THREAD🧵👇
2/ There will be 10 hours of different #VRChat world hops during #VeniceImmersive spanning 30 different worlds.
The guided tours are only available on site, but the worlds are public.
I link to all of the @VRChat worlds by genre at the bottom of this post: voicesofvr.com/1119-venice-im…
3/ The video below has trailers for the 30+ different #VRChat worlds featured in the #VeniceImmersive World Gallery Selection.