Tianyuan Zhang (@tianyuanzhang99) 's Twitter Profile
Tianyuan Zhang

@tianyuanzhang99

PhD students in MIT, working on vision. M.S. in CMU, B.S. in PKU

ID: 905014077126213632

linkhttp://tianyuanzhang.com calendar_today05-09-2017 10:25:45

77 Tweet

750 Followers

629 Following

Tianyuan Zhang (@tianyuanzhang99) 's Twitter Profile Photo

Really impressive results. I think data driven approaches will be able to do fully inverse/forward rendering soon, including strong specular effects, hard shadows and transparencies.

Yilun Xu (@xuyilun2) 's Twitter Profile Photo

Introducing Discrete-Continuous Latent Variable Diffusion Models (DisCo-Diff 🕺), which augment continuous diffusion models with learnable global discrete latents. DisCo-Diff greatly simplify learning diffusion models and strengthens their sampling trajectories (1/9)

Introducing Discrete-Continuous Latent Variable Diffusion Models (DisCo-Diff 🕺), which augment continuous diffusion models with learnable global discrete latents. DisCo-Diff greatly simplify learning diffusion models and strengthens their sampling trajectories (1/9)
Boyuan Chen (@boyuanchen0) 's Twitter Profile Photo

Diffusion Forcing Update: code & ckpt for 3D-UNet + Temporal Attention version is released thanks to my amazing undergrad mentee Kiwhan Song! See project github for more info. I also added suggested future directions to the our website boyuan.space/diffusion-forc…. Check them out!

Tianyuan Zhang (@tianyuanzhang99) 's Twitter Profile Photo

Got a chance to play pingpong in VR with this virtual agent on May, it’s so cool! Imagine more sophisticated interactions with virtual agent in the future.

S. Lester Li (@sizhe_lester_li) 's Twitter Profile Photo

Want to turn an IKEA lamp into an intelligent robot capable of learning and adapting autonomously by simply adding cameras and motors? Excited to share our multi-disciplinary project that blends robotics, computer vision, and next-gen fabrication to make it possible!

Ayush Tewari (@_atewari) 's Twitter Profile Photo

Excited to announce that I will be joining the University of Cambridge Engineering Dept as an assistant professor in spring 2025! I will be looking for students for the next year. Check out Elliott / Shangzhe Wu's thread for details on how to apply, and get in touch!

Excited to announce that I will be joining the University of Cambridge <a href="/Cambridge_Eng/">Engineering Dept</a> as an assistant professor in spring 2025! 

I will be looking for students for the next year. Check out <a href="/elliottszwu/">Elliott / Shangzhe Wu</a>'s thread for details on how to apply, and get in touch!
Tianwei Yin (@tianweiy) 's Twitter Profile Photo

Super excited to see Adobe Firefly’s Fast Mode out in the wild! 🎨 Thrilled to have contributed to the research that helped make this a reality. Can’t wait to see how creators unleash their imagination with it! 🚀 #AdobeFirefly

PyTorch (@pytorch) 's Twitter Profile Photo

Join us next Friday on October 11th at 10 AM PT for our next LIVE PyTorch Expert Exchange Webinar on Efficient Streaming Language Models with Attention Sinks w/ Guangxuan Xiao, MIT EECS ▶️🎙️ Tune in at: hubs.la/Q02SfV5v0 #LLMs

Join us next Friday on October 11th at 10 AM PT for our next LIVE PyTorch Expert Exchange Webinar on Efficient Streaming Language Models with Attention Sinks w/ <a href="/Guangxuan_Xiao/">Guangxuan Xiao</a>, <a href="/MITEECS/">MIT EECS</a> ▶️🎙️ Tune in at: hubs.la/Q02SfV5v0 #LLMs
Yilun Du (@du_yilun) 's Twitter Profile Photo

I'm recruiting PhD students this year with interest in machine learning, embodied AI, or AI for science! If you are interested in constructing fundamental tools to improve Generative AI and exploring how these tools can be used for intelligent embodied agents and science,

Yunzhi Zhang (@zhang_yunzhi) 's Twitter Profile Photo

Accurate and controllable scene generation has been difficult with natural language alone. You instead need a language for scenes. Introducing the Scene Language — a visual representation for high-quality 3D/4D generation by integrating programs, words, and embeddings — 🧵(1/6)

Haian Jin (@haian_jin) 's Twitter Profile Photo

Novel view synthesis has long been a core challenge in 3D vision. But how much 3D inductive bias is truly needed? —Surprisingly, very little! Introducing "LVSM: A Large View Synthesis Model with Minimal 3D Inductive Bias"—a fully transformer-based approach that enables scalable,

Tianyuan Zhang (@tianyuanzhang99) 's Twitter Profile Photo

Check our learning-based deterministic model for novel view synthesis without NeRF and 3DGS, yet producing consistent rendering(still need camera poses). We try to use few 3D inductive biases and make it simple!

Shivam Duggal (@shivamduggal4) 's Twitter Profile Photo

Current vision systems use fixed-length representations for all images. In contrast, human intelligence or LLMs (eg: OpenAI o1) adjust compute budgets based on the input. Since different images demand diff. processing & memory, how can we enable vision systems to be adaptive ? 🧵

Bowen Li (@bw_li1024) 's Twitter Profile Photo

Humans can learn to reason in an "unfamiliar" world, like new games. How far are LLMs from this? Check out our recent work @NeurIPS2024 D&B Track: "LogiCity: Advancing Neuro-Symbolic AI with Abstract Urban Simulation". Page: jaraxxus-me.github.io/LogiCity/