So Yeon (Tiffany) Min on Industry Job Market (@soyeontiffmin) 's Twitter Profile
So Yeon (Tiffany) Min on Industry Job Market

@soyeontiffmin

5th year PhD student at CMU MLD @mldcmu & Apple AI ML fellowship. Prev: Apple, Meta, B.S. and M.Eng from @MITEECS Advised by @rsalakhu and @ybisk.

ID: 1449394298294837255

linkhttps://soyeonm.github.io/ calendar_today16-10-2021 15:18:34

453 Tweet

755 Takipçi

240 Takip Edilen

Pratyush Maini (@pratyushmaini) 's Twitter Profile Photo

1/Pretraining is hitting a data wall; scaling raw web data alone leads to diminishing returns. Today DatologyAI shares BeyondWeb, our synthetic data approach & all the learnings from scaling it to trillions of tokens🧑🏼‍🍳 - 3B LLMs beat 8B models🚀 - Pareto frontier for performance

1/Pretraining is hitting a data wall; scaling raw web data alone leads to diminishing returns. Today <a href="/datologyai/">DatologyAI</a> shares BeyondWeb, our synthetic data approach &amp; all the learnings from scaling it to trillions of tokens🧑🏼‍🍳
- 3B LLMs beat 8B models🚀
- Pareto frontier for performance
Stephanie Milani (@steph_milani) 's Twitter Profile Photo

🌻 Excited to announce that I’ve moved to NYC to start as an Assistant Prof/Faculty Fellow at New York University! If you’re in the area, reach out & let’s chat! Would love coffee & tea recs as well 🍵

🌻 Excited to announce that I’ve moved to NYC to start as an Assistant Prof/Faculty Fellow at <a href="/nyuniversity/">New York University</a>! 

If you’re in the area, reach out &amp; let’s chat! Would love coffee &amp; tea recs as well 🍵
Kashu Yamazaki (@kashu_yamazaki) 's Twitter Profile Photo

この度、Forbes JAPANが選ぶ「世界を変える30歳未満の30人」に選出いただきました! これからも精進して研究します。日本を再びロボットの中心地に!!!!! 30 UNDER 30【ForbesJAPAN】 #u30fj

この度、Forbes JAPANが選ぶ「世界を変える30歳未満の30人」に選出いただきました!

これからも精進して研究します。日本を再びロボットの中心地に!!!!!

<a href="/forbesjapan_30/">30 UNDER 30【ForbesJAPAN】</a> #u30fj
Paul Liang (@pliang279) 's Twitter Profile Photo

A bit late, but finally got around to posting the recorded and edited lecture videos for the **How to AI (Almost) Anything** course I taught at MIT in spring 2025. Youtube playlist: youtube.com/watch?v=0MYt0u… Course website and materials: mit-mi.github.io/how2ai-course/… Today's AI can be

A bit late, but finally got around to posting the recorded and edited lecture videos for the **How to AI (Almost) Anything** course I taught at MIT in spring 2025.

Youtube playlist: youtube.com/watch?v=0MYt0u…

Course website and materials: mit-mi.github.io/how2ai-course/…

Today's AI can be
Chris Paxton (@chris_j_paxton) 's Twitter Profile Photo

Training a Whole-Body Control Foundation Model -- new work from my team at Agility Robotics A neural network for controlling our humanoid robots which is robust to disturbances, can handle heavy objects, and is a powerful platform for learning new whole-body skills learn more

Qian Huang (@qhwang3) 's Twitter Profile Photo

Yesterday was my last day at xAI. It’s been an incredible ride for the past year and half, probably the most adventurous and fast growing period of my life so far. Best wishes for the team going forward. Looking forward to what’s next!

Yesterday was my last day at <a href="/xai/">xAI</a>. It’s been an incredible ride for the past year and half, probably the most adventurous and fast growing period of my life so far. Best wishes for the team going forward. Looking forward to what’s next!
Sukjun (June) Hwang (@sukjun_hwang) 's Twitter Profile Photo

Coming from a computer vision background and now in sequence modeling, I’m often struck by how disconnected LLMs and vision feel. Our work, AUSM, treats video as language -- and it reveals a few blind spots we’ve overlooked.

Sachin Goyal (@goyalsachin007) 's Twitter Profile Photo

1/Excited to share the first in a series of my research updates on LLM pretraining🚀. Our new work shows *distilled pretraining*—increasingly used to train deployable models—has trade-offs: ✅ Boosts test-time scaling ⚠️ Weakens in-context learning ✨ Needs tailored data curation

1/Excited to share the first in a series of my research updates on LLM pretraining🚀.
Our new work shows *distilled pretraining*—increasingly used to train deployable models—has trade-offs:
✅ Boosts test-time scaling
⚠️ Weakens in-context learning
✨ Needs tailored data curation
Mahi Shafiullah 🏠🤖 (@notmahi) 's Twitter Profile Photo

Deeply honored to be a part of MIT Tech Review Innovators Under 35 List this year. This recognition highlights our work on building robot intelligence that generalizes to unseen and unstructured human environments, executed with my friends & colleagues at NYU Courant & beyond.

Deeply honored to be a part of MIT Tech Review Innovators Under 35 List this year.
This recognition highlights our work on building robot intelligence that generalizes to unseen and unstructured human environments, executed with my friends &amp; colleagues at NYU Courant &amp; beyond.
Sanket Vaibhav Mehta, Ph.D. (@sanketvmehta) 's Twitter Profile Photo

Mind-blowing view of #AuroraBorealis on my flight from JFK—>SFO. The wildest part? I edited this shot on my phone at 35000 ft using freakn Nano Banana 🍌 via Google Gemini App ♊️ and posting this from the same altitude right now thanks to Delta WiFi. The future is officially here ✨

Mind-blowing view of #AuroraBorealis on my flight from JFK—&gt;SFO. The wildest part? I edited this shot on my phone at 35000 ft using freakn <a href="/NanoBanana/">Nano Banana</a> 🍌 via <a href="/GeminiApp/">Google Gemini App</a> ♊️ and posting this from the same altitude right now thanks to <a href="/Delta/">Delta</a> WiFi. The future is officially here ✨
Dylan Sam (@dylanjsam) 's Twitter Profile Photo

🚨Excited to introduce a major development in building safer language models: Safety Pretraining! Instead of post-hoc alignment, we take a step back and embed safety directly into pretraining. 🧵(1/n)

🚨Excited to introduce a major development in building safer language models: Safety Pretraining!

Instead of post-hoc alignment, we take a step back and embed safety directly into pretraining.

🧵(1/n)
Pratyush Maini (@pratyushmaini) 's Twitter Profile Photo

We can’t keep slapping alignment bandage 🩹 on harmful LLMs and calling it safety. Let’s fix the leak at the source by making LLMs safe by design. Introducing, Safety Pretraining🛡️

We can’t keep slapping alignment bandage 🩹 on harmful LLMs and calling it safety. Let’s fix the leak at the source by making LLMs safe by design. 

Introducing, Safety Pretraining🛡️
Pratyush Maini (@pratyushmaini) 's Twitter Profile Photo

One thing years of memorization research has made clear: unlearning is fundamentally hard. Neurons are polysemantic & concepts are massively distributed. There’s no clean 'delete'. We need architectures that are "unlearnable by design". Introducing, Memorization Sinks 🛁⬇️

niki parmar (@nikiparmar09) 's Twitter Profile Photo

We just dropped Sonnet 4.5, the best coding model! Agents are truly here now -- autonomous task solving, complex multi-step tasks, parallel agents, combined with new tools and features and a lot more.. Check it out here 👇

Emily Byun (@yewonbyun_) 's Twitter Profile Photo

💡Can we trust synthetic data for statistical inference? We show that synthetic data (e.g. LLM simulations) can significantly improve the performance of inference tasks. The key intuition lies in the interactions between the moments of synthetic data and those of real data

💡Can we trust synthetic data for statistical inference?

We show that synthetic data (e.g. LLM simulations) can significantly improve the performance of inference tasks. The key intuition lies in the interactions between the moments of synthetic data and those of real data
Dylan Sam (@dylanjsam) 's Twitter Profile Photo

Very interesting insights into understanding when and why synthetic data (although imperfect and biased) can boost the performance of statistical inference!! 📈📈

Shrimai (@shrimai_) 's Twitter Profile Photo

Thank you Rohan Paul for highlighting our work!💫 Front-Loading Reasoning shows that inclusion of reasoning data in pretraining is beneficial, does not lead to overfitting after SFT, & has latent effect unlocked by SFT! Paper: arxiv.org/abs/2510.03264 Blog:

Jimin Mun (@jiminmun_) 's Twitter Profile Photo

Amazing theoretical work on how to generate text-based synthetic data that will *actually* improve performance on statistical inference!! 🤩