Ming-Yu Liu (@liu_mingyu) 's Twitter Profile
Ming-Yu Liu

@liu_mingyu

Tweets are my own.

ID: 4475055297

linkhttps://mingyuliu.net/ calendar_today13-12-2015 23:23:25

959 Tweet

8,8K Takipçi

489 Takip Edilen

Yin Cui (@yincuicv) 's Twitter Profile Photo

We released Cosmos-Reason1 code, model, and part of the data! We also updated our paper to include a section about our RL infra: arxiv.org/abs/2503.15558 - Code: github.com/nvidia-cosmos/… - Model and Data: huggingface.co/collections/nv… - Blog: developer.nvidia.com/blog/curating-…

Kaiwen Zheng (@zkwthu) 's Twitter Profile Photo

1/💡New paper from NVIDIA&Tsinghua International Conference on Minority Languages Spotlight! Direct Discriminative Optimization (DDO) enables GAN-style finetuning of diffusion/autoregressive models without extra networks. SOTA achieved on ImageNet-512! Website: research.nvidia.com/labs/dir/ddo/ Code: github.com/NVlabs/DDO

1/💡New paper from NVIDIA&amp;Tsinghua <a href="/ICML2025/">International Conference on Minority Languages</a> Spotlight!
Direct Discriminative Optimization (DDO) enables GAN-style finetuning of diffusion/autoregressive models without extra networks. SOTA achieved on ImageNet-512!
Website: research.nvidia.com/labs/dir/ddo/
Code: github.com/NVlabs/DDO
Max Zhaoshuo Li 李赵硕 (@mli0603) 's Twitter Profile Photo

Cosmos-Reason1 has exciting updates 💡 Now it understands physical reality — judging videos as real or fake! Check out the resources👇 Paper: arxiv.org/abs/2503.15558 Huggingface: huggingface.co/nvidia/Cosmos-… Code: github.com/nvidia-cosmos/… Project page: research.nvidia.com/labs/dir/cosmo… (1/n)

Ming-Yu Liu (@liu_mingyu) 's Twitter Profile Photo

We post-trained a reasoning model to reason whether a video is real or generated. It might be very useful as a critic to improve video generators. Take a look. NVIDIA AI

Ming-Yu Liu (@liu_mingyu) 's Twitter Profile Photo

For people looking for a diffusion-based video generator to finetune or post-train for their downstream physical AI applications, we just released our latest one. We have 2 models: 2B and 14B. 2B for fast prototyping and 14B for better quality. The license is fully open. Give it

kiui (@ashawkey3) 's Twitter Profile Photo

Happy to share our work PartPacker: We enable one-shot image-to-3D generation with any number of parts! Project page: research.nvidia.com/labs/dir/partp… Demo: huggingface.co/spaces/nvidia/… Code: github.com/NVlabs/PartPac…

Tsung-Yi Lin (@tsungyilincv) 's Twitter Profile Photo

Generating 3D models with parts is a key step toward scalable, interactive simulation environments. Check out our work — PartPacker — and the concurrent project, PartCrafter!" PartPacker: github.com/NVlabs/PartPac… PartCrafter: wgsxm.github.io/projects/partc…

Ming-Yu Liu (@liu_mingyu) 's Twitter Profile Photo

3D asset generation has advanced a lot in the past few years. Generating a holistic 3D asset is no longer a challenging problem. What's next for 3D generation? We believe that generating a 3D asset with individual parts defined is the next frontier. With the parts, we can start

Ming-Yu Liu (@liu_mingyu) 's Twitter Profile Photo

Big congrats to Eric Jang and the team on the 1X World Model release. Verification is an important part of producing production AI model. Given the diverse nature of the work environment, it makes a lot of sense to leverage a world model to help with policy evaluation.

Hanzi Mao (@hanna_mao) 's Twitter Profile Photo

We build Cosmos-Predict2 as a world foundation model for Physical AI builders — fully open and adaptable. Post-train it for specialized tasks or different output types. Available in multiple sizes, resolutions, and frame rates. 📷 Watch the repo walkthrough

Sean Kirmani (@seankirmani) 's Twitter Profile Photo

🤖🌎 We are organizing a workshop on Robotics World Modeling at Conference on Robot Learning 2025! We have an excellent group of speakers and panelists, and are inviting you to submit your papers with a July 13 deadline. Website: robot-world-modeling.github.io

🤖🌎 We are organizing a workshop on Robotics World Modeling at <a href="/corl_conf/">Conference on Robot Learning</a> 2025!

We have an excellent group of speakers and panelists, and are inviting you to submit your papers with a July 13 deadline.

Website: robot-world-modeling.github.io
Daniel Ho (@itsdanielho) 's Twitter Profile Photo

We at 1X with Jack Monas are excited to announce the ICCV phase of our 1X World Model Challenge: huggingface.co/spaces/1x-tech… Participate in the Compression and Sampling tracks for a $8k prize pool & train generative models for cool robot results like: 1x.tech/discover/redwo…

Ming-Yu Liu (@liu_mingyu) 's Twitter Profile Photo

Together with Aaron Lefohn and Sanja Fidler, we will give a special address at SIGGRAPH. Specifically, I will give an update on our vision and our current work in enabling Physical AI. Please join us. nvidia.com/en-us/events/s…

NVIDIA Omniverse (@nvidiaomniverse) 's Twitter Profile Photo

Kick off your #OpenUSD Day with a look into the future of robotics and autonomous vehicles. 🤖 Join Ming-Yu Liu as he shares how #NVIDIACosmos world foundation models unlock prediction and reasoning for the next wave of robotics and autonomous vehicles. 📅Wednesday, 8/13 at

Ming-Yu Liu (@liu_mingyu) 's Twitter Profile Photo

In Cosmos, we are hiring Cosmos World Foundation Model builders. If you are interestd in building large-scale video foundaiton model and multimodal LLM for Robots and cars, please send your CV to [email protected] If you have experiences in large-scale diffusion models,

Ming-Yu Liu (@liu_mingyu) 's Twitter Profile Photo

The submissions portal for the NVIDIA 2026-2027 Graduate Fellowships is now open research.nvidia.com/graduate-fello… PHD students work on AI. Please apply!

Jiahui Huang (@huangjh_hjh) 's Twitter Profile Photo

[1/N] 🎥 We've made available a powerful spatial AI tool named ViPE: Video Pose Engine, to recover camera motion, intrinsics, and dense metric depth from casual videos! Running at 3–5 FPS, ViPE handles cinematic shots, dashcams, and even 360° panoramas. 🔗 research.nvidia.com/labs/toronto-a…