LLaMA spearheaded the age of community-driven midsize LMs. It’s game-changing for many use cases that don’t require GPT-4’s full firepower.
Great blog post that covers Alpaca, Vicuna, Koala, WizardLM, OpenAssistant, and more!
An “AI detector” incorrectly flagged a student’s essay as AI-generated.
Banning AI systems/ trying to detect AI generated content in school is a losing battle. The purpose of education is to prepare students for the world and economy of tomorrow. That world will include AI.
Exciting news! I am pleased to announce that the code for the Dextreme project in Isaac Gym is now available on GitHub: https://github.com/NVIDIA-Omniverse/IsaacGymEnvs… As far as I am aware, this is the first publicly accessible GPU-accelerated ADR (automatic domain randomization) implementation, and it…Show more
DeXtreme is our new work on scaling sim-to-real for contact-rich manipulation with a vision-based state estimation on a robot hand with the infrastructure we have been developing with Isaac Gym over the past one year.
https://arxiv.org/abs/2210.13702https://dextreme.org
We are looking at the future of VR, YouTube & Google Street View.
This is zip-NeRF, a 3D neural rendering tech rapidly approaching the quality of a real, high-res drone flight video. Think of NeRF as transporting reality into simulation. Metaverse will finally work this time.
BREAKING
OpenAI released a implementation of Consistency Models
consistency models, a new family of generative models that achieve high sample quality without adversarial training. They support fast one-step generation by design, while still allowing for few-step sampling to…Show more
Its over
run Large Language Models like LLaMA, llama.cpp, GPT-J, Pythia, OPT, GALACTICA, gpt4all, auto-gpt easily in a web ui, free, and open source
github:
How does a Large Language Model like ChatGPT actually work?
Well, they are both amazingly simple and exceedingly complex at the same time.
Hold on to your butts, this is a deep dive ↓
.
We combine guided trajectory diffusion with a physics-based humanoid controller to enable pedestrian animation that is controllable by a user.
Project page: https://nv-tlabs.github.io/trace-pace/
1/5
Today we're releasing the Segment Anything Model (SAM) — a step toward the first foundation model for image segmentation.
SAM is capable of one-click segmentation of any object from any photo or video + zero-shot transfer to other segmentation tasks ➡️https://bit.ly/433YuBI
We are proud to introduce Wonder Studio.
An AI tool that automatically animates, lights and composes CG characters into a live-action scene. No complicated 3D software, no expensive production hardware—all you need is a camera.
Sign up for closed beta at http://wonderdynamics.com
AI tool can detect Photoshop alterations and produce a heat map of the areas on the subject's face that have been edited.
Paper: https://peterwang512.github.io/FALdetector/
Before/after of Corridor's latest AI video is wild. They shot video on greenscreen, made virtual sets in Unreal, then reskinned it to anime by finetuning Stable Diffusion. Net result? 120 VFX shots done by a team of 3 on a dime. Bravo! This is a milestone in creative technology🧵
Almost 10 years ago, researchers at ETH caused a sensation with a robot cube that could balance independently balancing on its pivot. Now its successor has arrived: the One-Wheel Cubli. It only needs one reaction wheel.🤖#Robotics
Amazing! text-to-3D including anims! 🤯
A text prompt input is turned into a good looking 3D model which is also rigged and can be animated with another prompt.
Democratization of 3D content creation.🚀
Very impressive what
Society needs to “consider the best ways for doctors and AI to work together, and how medical roles will change as a consequence”, argue @EffyVayena and @ProfADMorris in a guest essay https://econ.st/3EKYI6l
If you liked ICON, give ECON a try. It gives better, more detailed, 3D human shapes from a single image with all the same benefits of ICON. Works for varied poses and has #SMPL under the hood. Congratulations
ECON got accepted by #CVPR2023
Detailed clothed human recovery from single image via normal integration.
Is implicit MLP a must? NO.
Is data-driven/learning a must? NO.
How to keep pose robustness w/o sacrificing the topological flexibility?
See https://xiuyuliang.cn/econ
I finally find an explanation for why RL is needed for RLHF that satisfied me. It's actually like playing board games.
The reward model can only judge a full answer and a "critic" is needed to efficiently improve the intermediate moves (earlier tokens in the answer) 1/4
🎉🎉 Our paper "Next3D: Generative Neural Texture Rasterization for 3D-Aware Head Avatars" has been accepted to present at #CVPR2023!
We Introduce a 3D GAN for high-quality facial avatars and welcome to check out its applications on one-shot 3D animation and stylization😎
Language Is Not All You Need: Aligning Perception with Language Models
introduce KOSMOS-12, a Multimodal Large Language Model (MLLM) that can perceive
general modalities, learn in context (i.e., few-shot), and follow instructions (i.e., zero-shot)
abs: https://arxiv.org/abs/2302.14045
! I am blown away by the AI video generating tool's ability to provide endless possibilities for the future of VFX, animation, storytelling, & content production.