Haomeng Zhang (@haomengz99) 's Twitter Profile
Haomeng Zhang

@haomengz99

Ph.D. @PurdueCS working on Computer Vision & Robotics; @IllinoisCS @UMichCSE @sjtu1896 Alumni

ID: 1238672762547449857

linkhttps://haomengz.github.io/ calendar_today14-03-2020 03:46:26

49 Tweet

387 Takipçi

966 Takip Edilen

OpenAI (@openai) 's Twitter Profile Photo

Our new text-to-image model, DALL·E 3, can translate nuanced requests into extremely detailed and accurate images. Coming soon to ChatGPT Plus & Enterprise, which can help you craft amazing prompts to bring your ideas to life: openai.com/dall-e-3

Our new text-to-image model, DALL·E 3, can translate nuanced requests into extremely detailed and accurate images.

Coming soon to ChatGPT Plus & Enterprise, which can help you craft amazing prompts to bring your ideas to life:

openai.com/dall-e-3
Sirui Xu (@xu_sirui) 's Twitter Profile Photo

Can we use a diffusion model to generate various everyday interactions between the whole body and dynamic objects?🧐 🔥Delighted to share our work InterDiff #ICCV2023🔥 Project: sirui-xu.github.io/InterDiff Code: github.com/Sirui-Xu/Inter… Arxiv: arxiv.org/abs/2308.16905 🧵[1/6]

Angela Dai (@angelaqdai) 's Twitter Profile Photo

We've released the ScanNet++ data! Check it out: kaldir.vc.in.tum.de/scannetpp/ 280 high-fidelity 3D scenes w/ 1mm geometry, DSLR+iPhone images, semantics We're currently beta-testing, please bear with us - approval may initially take up to 2 weeks Test scenes and benchmark to come!

We've released the ScanNet++ data!
Check it out: kaldir.vc.in.tum.de/scannetpp/
280 high-fidelity 3D scenes w/ 1mm geometry, DSLR+iPhone images, semantics

We're currently beta-testing, please bear with us - approval may initially take up to 2 weeks

Test scenes and benchmark to come!
Yixuan Wang (@yxwangbot) 's Twitter Profile Photo

What should the right representation for robotic manipulation be? Enter D^3Fields: a 3D, dynamic, and semantic representation using foundation models WITHOUT training for zero-shot generalizable robotic manipulation. Colab is available! 🔗 robopil.github.io/d3fields/ 🧵👇

OpenAI (@openai) 's Twitter Profile Photo

Introducing Sora, our text-to-video model. Sora can create videos of up to 60 seconds featuring highly detailed scenes, complex camera motion, and multiple characters with vibrant emotions. openai.com/sora Prompt: “Beautiful, snowy

Md Ashiqur Rahman (@ashiq_rahman_s) 's Twitter Profile Photo

Excited to present our work, CoDA-NO, on multi-physics systems! Join us at the first poster session of #NeurIPS2024 on Wednesday, December 11, in the East Exhibit Hall, Poster #4102. See you there!

Freda Shi (@fredahshi) 's Twitter Profile Photo

I received a review like this five years ago. It’s probably the right time now to share it with everyone who wrote or got random discouraging reviews from ICML/ACL.

I received a review like this five years ago. It’s probably the right time now to share it with everyone who wrote or got random discouraging reviews from ICML/ACL.
Md Ashiqur Rahman (@ashiq_rahman_s) 's Twitter Profile Photo

Our paper "Group Downsampling with Equivariant Anti-aliasing" will be presented at #ICLR2025 🎉! We propose a novel subgroup sampling layer connecting Cayley graphs, uniform subgroup subsampling, and anti-aliasing—boosting equivariant models' efficiency with minimal compute.

Our paper "Group Downsampling with Equivariant Anti-aliasing" will be presented at #ICLR2025 🎉!

We propose a novel subgroup sampling layer connecting Cayley graphs, uniform subgroup subsampling, and anti-aliasing—boosting equivariant models' efficiency with minimal compute.
Martin Ziqiao Ma (@ziqiao_ma) 's Twitter Profile Photo

Vision-Language Models (VLMs) can describe the environment, but can they refer within it? Our findings reveal a critical gap: VLMs fall short of pragmatic optimality. We identify 3 key failures of pragmatic competence in referring expression generation with VLMs: (1) cannot

Amber Yijia Zheng (@amberyzheng) 's Twitter Profile Photo

🚨 We are presenting our ICML 2025 oral next week: Model Immunization from a Condition Number Perspective 📄 amberyzheng.com/immu_cond_num/ (1/n)

🚨 We are presenting our ICML 2025 oral next week: Model Immunization from a Condition Number Perspective
📄 amberyzheng.com/immu_cond_num/
(1/n)
Raymond A. Yeh (@raymondyeh) 's Twitter Profile Photo

Tomorrow, we are presenting “Model Immunization from a Condition Number Perspective” at ICML: 📢Oral: Jul 17, 1:45–2:00 p.m. EDT @ West Exhib. Hall C 📌Poster: 2:00–4:30 p.m. EDT @ East Exhib. Hall A-B (E-1604) Come talk to Cedar and learn more about reducing model misuse!

Jiraphon Yenphraphai (@jyenphraphai) 's Twitter Profile Photo

[1/3] 🚀 Introducing ShapeGen4D: video → high-quality 4D mesh sequences. A native, end-to-end video-to-4D model that turns monocular videos into high-quality mesh sequences (no per-frame optimization). details 👉 shapegen4d.github.io

Md Ashiqur Rahman (@ashiq_rahman_s) 's Twitter Profile Photo

Excited to present our work “Local Scale Equivariance with Latent Deep Equilibrium Canonicalizer” at #ICCV2025! 🌺 We introduce DEC, a plug-and-play module that enhances local scale equivariance in vision models like ViT and Swin, boosting robustness with minimal overhead.

Excited to present our work “Local Scale Equivariance with Latent Deep Equilibrium Canonicalizer” at #ICCV2025! 🌺
We introduce DEC, a plug-and-play module that enhances local scale equivariance in vision models like ViT and Swin, boosting robustness with minimal overhead.