LLMs like those used by ChatGPT are missing a crucial language for faithfully interpreting the physical world, without which, they have a ceiling for what questions can be answered/predicted.
LLMs have no physical intuition as they are trained exclusively on text.
They may correctly answer physical intuition questions *if* they can retrieve answers to similar questions from their vast associative memory.
But they may get the answer *completely* wrong
1/
One of the hardest challenges in developing AI for autonomous vehicles is evaluating the performance of our driving models. Why? (A short 🧵on our latest research on multi-agent RL).
It looks like Gordon Bell's 1998 vision of 'MyLifeBits' has found its way into the world. I believe this will be a significant theme over the next decade https://en.wikipedia.org/wiki/MyLifeBits
Introducing Rewind
Today we are unveiling Rewind: the search engine for your life. It’s a macOS app that enables you to find anything you’ve seen, said, or heard.
Here’s how it works:
(5/23)
https://youtube.com/watch?v=dIV0ZiZluQo…
Authenticated identity is a necessity not a premium feature for many web3 applications. Seems Elon is in a rare situation where tying identity to account is actually desired by a lot of users. Will be interesting to see how this plays out.
Fun fact: in 2015 I registered the domain name http://MetaverseFestival.com as I thought there was going to be a future where immersive virtual worlds could be traversed with ease, and that one day it would be cool to host a global virtual festival to celebrate humanity's diversity 🙃
Cognitive Assistance Lab for your work to enable audio wayfinding for people who are visually impaired, announced at Connect today.
Extended video of the partnership here, featuring yours truly 😊https://facebook.com/TechatMeta/videos/1180323275899569/…
Excellent turnout for our tutorial on egocentric vision research using #projectaria at #CVPR2022. Learn about how Project Aria is being used to advance AI & ML research, including concretizing privacy challenges for machine perception in AR - agenda @ https://ariatutorial2022.github.io
In addition to “Daily Activities”, the dataset also contains 16 sequences of “Desktop Activities” captured with a multi-cam setup, for further accelerating research into human-object interactions.
7/n
In addition to sensor data, the Pilot Dataset also contains results from machine perception services which provide additional context to the spatial-temporal reference frames, such as:
Accurate 3D trajectories of users across multiple everyday activities in the same location
3/n
The Project Aria Pilot Dataset consists of 143 sequences of “Daily activities”, consisting of over 7 accumulated hours of data captured on Aria with additional face blurring to demonstrate our commitment to responsible innovation.
2/n
Ahead of #CVPR2022 , I’m excited to share the open dataset of Project Aria data from Meta Reality Labs, along with accompanying open research tools designed to accelerate AI and ML research. https://about.facebook.com/realitylabs/projectaria/datasets…
A little about the dataset and why I think it’s so exciting…
1/n
This @clipdropapp × @figma plugin is truly mindblowing
It can automatically decompose images into multiple layers. No more issues with images not fitting a layout
Let me know if you'd like to try it
cc @zoink#FigmaCommunity#Config2022
With our new measurement methodology, we’ve halved the testing time needed to obtain robust insight.
Find out how we're using the latest methods in data science and ML to accelerate the speed at which we can test and evaluate driving models on the road.
When we re-think interaction and use AR as a human-machine interface, a lot of new possibilities emerge. Get in touch if you have a use case you'd like to explore. Full video: https://youtu.be/2zxIPkpYseo#construction#manufacturing#factory#automation#robots#augmentedreality
How long before we have a truly live World Map to explore? How long after that can we jump into it virtually?
How long before we can use that map to code against reality, for AVs, for AR, for AI?
This idea will never cease to fascinate me! 🤯
.@KevinJames' son convinced him to get an Oculus in order to be more active and then Kevin found his son standing there ringing people up at a virtual convenience store. #FallonTonight
's progress over the last few years. From early explorations in end-to-end learning, to the first autonomous vehicle solution capable of scaling to 100 cities. Well done
As a…
– user
I want…
– an immersive VR shopping experience
so that…
– I can enjoy the inconvenience of physical shopping from the convenience of my home
& the team (11k paid subscribers grown organically) and I'm truly excited to see as the team roll out their API to turbo-charge the value to creators in the coming months. 3/4
will be a major player in that ecosystem, and I'm thrilled to be participating in their journey as the team strives to dramatically lower the barriers for creating high quality product visuals, powered by a touch of AI. 2/4
#2: Berlin is the place to be in Europe for AR creators! Would very much love to meetup in person some time and experience some of your AR demos again in reality!
#1: Spaces is super compelling. Different to video call, different to group text. I didn't use clubhouse but definitely see that group voice broadcasts warrant their own category.
Driving in London can be hard for anyone to do let alone a #selfdriving car. See how our AI-driven AVs are learning to navigate the busy streets of London using only computer vision, narrated by Wayve's CEO @alexgkendallhttps://youtu.be/h8AhWKOUSPU
Amazing results from @sonanticlabs, creating a text-to-speech model to revive Val Kilmer's voice for him, after he lost it to cancer. Could be an interesting partner to