Kent Bye (Voices of VR) Profile picture
Feb 23, 2022 42 tweets 35 min read Read on X
1/ Brief THREAD to unpack @MetaAI's new Project CAIRaoke AI self-supervised learning model announced this morning.

Launch video: facebook.com/zuck/videos/67…

Zuckerberg showed a demo of a Builder Bot conversational VR worldbuilding tool within Horizon Worlds that's featured below.
2/ Horizon World has terrible graphics, but Self-Supervised Learning is foundational AI tech that'll allow multi-modal training beyond labeled data, making speech-to-3D object creation in immersive worlds a reality. Will this yield good world design? TBD.
3/ Project CAIRaoke is a new conversational assistant synthesizing Natural Language Understanding, Dialog State Tracking, Dialog Policy, Management, & Natural Language Generation.
They say it's "deeply contextual & personal."
Video:
4/ This announcement is another incremental step towards @Meta's vision of "contextually-aware AI" that can only be achieved via "egocentric data capture" & "egocentric perception."
Essentially omniscient AI that can respond to your situation & context.
5/ Zuckerberg mentioned #Ego4D, which is 3.7k hrs of egocentric video data in collab w dozen universities to make AI training benchmarks around Episodic Memory, Hand-Object Interactions, AV Diarization, Social, & Forecasting.

More: ego4d-data.org
6/ I understand the utility of #Ego4D for AR & how it's the next frontier of Computer Vision + AI research, but it's still somewhat creepy as it's also in the context of @meta's omniscient Contextually-Aware AI in the service of surveillance capitalism.
7/ Contextually-Aware AI #Ego4D Benchmarks = Surveillance
Episodic Memory
Querying Memory
Query Construction
Recalling Lives
Hand-Object Interactions
State Changes
AV Diarization
Hearing Words
Tracking Conversations
Social Gaze
Social Dynamics
Forecasting Movements & Interactions
8/ The challenge with Contextually-Aware AI is ensuring it's aligned with Nissenbaum's Contextual Integrity Theory of Privacy ensuring proper flows of information.

How & why is this omniscient-level of data being used & to what ends?

My podcast interview
voicesofvr.com/998-primer-on-…
9/ I attended a Meta press briefing on Contextually-Aware AI on 3/18/21, & I paraphrased their goals along with some of my initial reactions to it in this podcast.

SPOILER: I'm not yet convinced of contextually-aware, omnipresent & omniscient AI overlords
voicesofvr.com/985-facebook-h…
10/ There's two other projects @Meta announced today.
1. No Language Left Behind: translation system that can learn any language.
2. Universal Speech Translator: real-time speech-to-speech translation across all languages(!!!) which sounds amazing (but never get to 100% accuracy)
11/ Zuckerberg emphasized that some of research that Meta is doing with AI is fundamental as it is being used in other contexts like speaking up MRIs up to 4x:
12/ Zuckerberg claims Meta is commited to building openly & responsibly as AI technologies "deliver the highest levels of privacy & help prevent harm" citing CrypTen.
But not sure how contextually-aware AI will deliver "the highest level of privacy."
It's the opposite of that.
13/ That's the highlights from 15-min intro video Zuckerberg streamed from his page.
There's another 2.5 hour video "Inside the Lab: Building for the Metaverse with AI" on @MetaAI's page digging into more topics.

Schedule: ai.facebook.com/events/inside-…
Video: facebook.com/MetaAI/videos/…
14/ I watched this 2.5 hour program from Meta on "Building for the Metaverse with AI," and the target audience is potential AI engineers as it was an extended job recruitment pitch, especially at the end (see slides below).

But I'll share some reflections on AI tech & ethics.
15/ Evolution of @Meta's ML architectures is interesting
Past: Supervised Learning & Task-Specific Models
Present: Self-Supervised Learning on Task-Independent Models to Tasks, Product & Tech
Future: SSL with Unified Model & Complex Prediction Tasks
NOTE: Sentiment Analysis & Ads
16/ In Building for the Metaverse, @Meta will combine Self-Supervised Learning, Continual Learning, Reinforcement Learning, & Planning & Reasoning to ALL XR data like biometric sensor data + egocentric data + environmental data to make Contextually-Relevant AI & a "World" Model.
17/ This all encompassing "World" model is so important is because of the types of multi-modal sensor fusion that @Meta will be doing.
For example, at #IEEEVR 2021 @RealityLabs showed how they could extrapolate Eye Gaze from Hand Pose + Head Pose:
18/ Here's a taxonomy of how different XR data & biometric measurements can do psychographic profiling that tracks XR user's actions, mental thoughts, unconscious physiological reactions & emotional states.
What are the mental privacy implications of fusing all of these together?
19/ Human Rights Lawyer @brittanheller gave a great @Gatherverse speech on 5 things XR privacy folks should understand about how XR data is completely different including lack of legal definition of this data.

See my video on this:

20/ @Meta first published their Five Pillars of Responsible AI on June 22, 2021 & are diff than their four RI Principles:
Privacy & Security
Fairness & Inclusion
Robustness & Safety
Transparency & Control
Governance & Accountability
ai.facebook.com/blog/facebooks…
21/ In April 2021, @Meta published a Privacy Progress Update with 8 Core Principles of how they define privacy, which is all about controlling access to identified data, but there's nothing on physiological XR data or profiling via biometric psychography.
22/ @Meta is still in an old paradigm of defining privacy as identity & control of data rather than contextually-relevant psychographic profiling via biometric data.

They still haven't commented on #NeuroRights like a Right to Mental Privacy.

See my talk
23/ Meta's Five Pillars of Responsible AI is an incremental step to RI best practices, but there's still a long ways to go.

Two talks with more
Towards a Framework for XR Ethics


Sensemaking Frameworks for the Metaverse & XR Ethics
24/ Meta's consistently frames privacy issues back to individual identity & the ways they're protecting personal identifiable information from leaking out in their ML training analysis.
They also oddly often shifted from privacy into open source & transparency tools like CrypTen.
25/ Back to the overall thrust of the presentation, it covered major applications & intentions for @MetaAI's efforts, and how their research is being applied to the Metaverse, which I'll dig into a bit.
They focused primarily on Robotic Embodiment, Creative Apps, & Safety.
26/ Here's a number of graphics showing the faster improving performance of their Self-Supervised Learning approaches vs Supervised Learning with different algorithms progressing over the years.
27/ Here's a number of different haptics & robotics applications of Self-Supervised Learning.

[NOTE: Searching for Meta vs Facebook has killed their SEO in trying to track down the original papers or references of some of this]

Blog on SSL from 3/4/21:
ai.facebook.com/blog/self-supe…
28/ Here's a series of graphics that shows what modules within an Classical ConvAI Pipeline that Meta's new CAIRaoke Model replaces.
[NOTE: yes, it is pronounced "karaoke"]
Additionally, the CAIRaoke output is recursively fed back in as an iterative input with sample assertions.
29/ @Meta showed a speculative AR design demonstrating how creepy Contextually-Aware AI might be.
It knows you already added enough salt & automatically buys more salt for you.
Has a memo from your mom + shows how she sliced Habeneros thinly, unclear if AI is extrapolating these.
30/ @lexfridman facilitated a really engaging 40-min discussion with pioneering AI researchers Yoshua Bengio & @MetaAI's @ylecun on the pathway to human-level AI, self-supervised learning, & consciousness (Bengio likes Graziano's Attention Schema Theory).
facebook.com/MetaAI/videos/…
31/ There was a whole section on language translation that I found interesting. I know how speech-to-tech is not and may never reach 100% accuracy, & so I really wonder how to quantify the accuracy & communication loss of the vision of these universal translators.
32/ They used to use English as an intermediate language to translate into, but they're moving to a system where they can directly translate any language into another language.
It makes me wonder about gaps in contextual knowledge & how anyone will ever detect mistranslations
33/ Some slides describing how Self-Supervised Learning works by taking away information & then requiring the ML to learning abstract schemas to be used for sensemaking & filling in the gaps. This can be iterated on for SSL to learn how to discern patterns more efficiently.
34/ There was a whole section on Responsible AI approaches & projects that @Meta has been working on.
The process of Tracking Fairness Across Meta in AI is a bit of a more robust & fleshed out responsible innovation framework than Reality Labs has talked about so far.
35/ I don't fully understand how they can model & explain SSL models.
'AI Model Cards are a standardized way to document, track, & monitor individual models...'
'AI System Cards how a group of AI System of AI Models & other non-AI techniques work together to accomplish a task.'
36/ Part of my hesitation in diving into running everything on self-learning AI models is it's very easy to create complex systems that you don't fully understand.
I clicked on links quickly tracking down a story & I was auto-blocked by Meta's AI overlords
37/ There's a lot of ways in which both safety & security requirements for moderation for XR AND the challenges of immersive VR & AR content + egocentric data capture are all catalysts for new AI benchmarks & challenges and even sometimes new algorithmic approaches.
38/ Starting to wrap up this thread.
These types of talks by Meta are quite dense to sift through, but they also usually reveal quite a lot of interesting insights into their deeper strategies & philosophies.
Also VR, AR, & the Metaverse are great provocations for AI innovations.
39/ Ultimately, this was a recruitment pitch.
It's the most optimistic story that @Meta tells itself about who they are, what they do, & why.
Creating AI to benefit people & society.
Building the future responsibly.
Ensuring fairness & robustness.
Collaborating with stakeholders.
40/ I have so many disconnects when I hear Zuckerberg say he want to "deliver the highest levels of privacy" while also aspiring to create contextually-aware AI with #Ego4D that can reconstruct memories, track every conversation, & predict your next move.
41/ The XR industry is in a phase of self-regulating XR Ethics, & so it's worth focusing on what @Meta is saying & doing.

I have more critical analysis on Meta in my opening @Gatherverse Summit keynote here + my Gatherverse thread with more context 👇
END/ If you find value in this type of coverage of the XR industry, then please consider supporting me on Patreon.
Thanks!
patreon.com/voicesofvr

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Kent Bye (Voices of VR)

Kent Bye (Voices of VR) Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @kentbye

Oct 11, 2022
1/ THREAD covering @Meta's #MetaConnect, which is starting now.

I'll be covering all of the announcements, watching all new videos, & reading all of their press releases. They usually dump a ton of new info

Follow here: facebook.com/events/6307993…
Or here:
2/ Zuckerberg saying we're going to be taking another big step forward with VR after walking through a history of VR tech since the Oculus acquisition.
3/ @boztank talking about some ecosystem stats for VR apps.
Philosophically they want to make ecosystem more open & how some apps have onboarded via the App Lab. Will push towards openness.
Read 119 tweets
Sep 22, 2022
1/ @PICOXR officially announced the Pico 4 today.

Pico was acquired by Bytedance (aka @tiktok_us) & they’re taking on Meta as the best standalone VR competitor to the Quest (but only in Europe & Asia to start).

Thread 🧵👇 to digest the livestream & news
ImageImageImageImage
2/ My sources of @PICOXR news are the official livestream👆

@CasandChary have the most comprehensive & digest & hands on:


@vr_oasis did a hands-on analysis


@RtoVR has the best specs breakdown here:
roadtovr.com/pico-4-announc… Image
3/ Here's 1st of 4 trailers @PICOXR showed during their Pico 4 announcement livestream.

Pico 4 has color, mixed reality passthrough while Meta is expected to launch similar features with Meta Quest Pro but at a higher price point.
Pico 4 Pro details to be announced at AWE Lisbon
Read 33 tweets
Sep 14, 2022
1/ Starting now is the #ARVRPolicyConf22 led by @XRAssociation & @ITIFdc.


I'll be tuning in & covering the talks in thread🧵👇

I expect the framing will be business-friendly as XRA is a trade organization representing Meta, Unity, MSFT, Sony, & Google.
2/ @XRAssociation's Liz Hyman opens the #ARVRPolicyConf22 & mentions they just launch a 4-page #HowXRWorks infographic yesterday called “Into the Digital World: XRA’s Guide to Immersive Technology”
xra.org/howxrworks/
3/ Reality Caucus co-chair @repdarrellissa gives the opening keynote at #ARVRPolicyConf22.
delbene.house.gov/news/documents…

Says that they were able to include instructions for immersive technology for the National Science Foundation & insert AI language into the THUD bill.
Read 70 tweets
Sep 13, 2022
1/ I recorded an epic recap & reviews of the 30 immersive stories in competition at #VeniceImmersive with @vrgeschichten.

We chat about each experience for ~5 min, & so it's a comprehensive break down of this year's selection.
voicesofvr.com/1121-venice-im…

20+ interviews coming soon Image
2/ I'm about to release other 20 podcasts from #VeniceImmersive, and here's a sneak peak of the 26 episodes, which total just over 24 hours worth of coverage. Image
3/ Here's the Twitter thread of my live coverage from #VeniceImmersive, which starts with an interview with co-curators Liz Rosenthal and Michel Reilhac.
voicesofvr.com/1119-venice-im…
Read 28 tweets
Aug 29, 2022
1/ Venice Immersive features 75 immersive stories.
I got a preview of highlights by co-curators @LizRosenthal & @michelreilhac (includes links to 30+ #VRChat experiences)
voicesofvr.com/1119-venice-im…

I'll be onsite at #VeniceImmersive covering the fest & this is will be my THREAD🧵👇 Image
2/ There will be 10 hours of different #VRChat world hops during #VeniceImmersive spanning 30 different worlds.
The guided tours are only available on site, but the worlds are public.
I link to all of the @VRChat worlds by genre at the bottom of this post:
voicesofvr.com/1119-venice-im… Image
3/ The video below has trailers for the 30+ different #VRChat worlds featured in the #VeniceImmersive World Gallery Selection.

Find links to these worlds at the bottom of this post: voicesofvr.com/1119-venice-im…

VIDEO:
Read 46 tweets
Aug 23, 2022
Meta is rolling out their Meta accounts today.

They’re introducing a follower model option, which will be interesting to see how that works with a medium that is primarily synchronous.

Blog overview
oculus.com/blog/getting-s…

You can still use Oculus acct on Quest 1 until 2023. ImageImage
2/ I migrated my account and made it “Open to everyone” to follow + “Anyone can see this profile.”

Friends are now called “Mutuals” and folks who you’d don’t follow but follow you are followers who can track your activities.

See
store.facebook.com/help/accounts/…

store.facebook.com/help/quest/art… ImageImageImageImage
3/ The new language around Mutuals as Friends and “follow requests” is confusing.

A “follow request” should really be named “mutual follow request” as accepting a “follow request” is the same as accepting a friend request.

“Should I let this person follow me?”= friending them. Image
Read 6 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(