Dylan Sam (@dylanjsam) 's Twitter Profile
Dylan Sam

@dylanjsam

phd student @mldcmu | past: student researcher @GoogleAI, intern @AmazonScience, undergrad @BrownUniversity

ID: 921531290674958336

linkhttps://dsam99.github.io/ calendar_today21-10-2017 00:19:15

160 Tweet

768 Followers

439 Following

Paul Liang (@pliang279) 's Twitter Profile Photo

This spring I am teaching a new class at MIT called **How to AI (Almost) Anything** Its name is a play on 2 seminal MIT Media Lab courses: how to make almost anything (on design & fabrication) and how to grow almost anything (on synthetic biology) We are now in the AI age, and

This spring I am teaching a new class at MIT called **How to AI (Almost) Anything**

Its name is a play on 2 seminal <a href="/medialab/">MIT Media Lab</a> courses: how to make almost anything (on design &amp; fabrication) and how to grow almost anything (on synthetic biology)

We are now in the AI age, and
Mingjie Sun (@_mingjiesun) 's Twitter Profile Photo

What makes a Large Language Model unique? I am excited to share our new work “Idiosyncrasies in Large Language Models”. We demonstrate that LLMs exhibit idiosyncrasies – unique patterns in their outputs that enable us to distinguish these models with exceedingly high accuracies.

What makes a Large Language Model unique? I am excited to share our new work “Idiosyncrasies in Large Language Models”.

We demonstrate that LLMs exhibit idiosyncrasies – unique patterns in their outputs that enable us to distinguish these models with exceedingly high accuracies.
Fahim Tajwar (@fahimtajwar10) 's Twitter Profile Photo

Interacting with the external world and reacting based on outcomes are crucial capabilities of agentic systems, but existing LLMs’ ability to do so is limited. Introducing Paprika 🌶️, our work on making LLMs general decision makers than can solve new tasks zero-shot. 🧵 1/n

Interacting with the external world and reacting based on outcomes are crucial capabilities of agentic systems, but existing LLMs’ ability to do so is limited.

Introducing Paprika 🌶️, our work on making LLMs general decision makers than can solve new tasks zero-shot.

🧵 1/n
Pratyush Maini (@pratyushmaini) 's Twitter Profile Photo

1/Being in academia is such a privilege: You get to collaborate with insanely talented & passionate students on their journey to upskill themselves. Very excited to share *OpenUnlearning*: a unified, easily extensible framework for unlearning led by Anmol Mekala Vineeth🧵

1/Being in academia is such a privilege: You get to collaborate with insanely talented &amp; passionate students on their journey to upskill themselves.

Very excited to share *OpenUnlearning*: a unified, easily extensible framework for unlearning led by <a href="/anmol_mekala/">Anmol Mekala</a> <a href="/VineethDorna/">Vineeth</a>🧵
Nicholas Roberts (@nick11roberts) 's Twitter Profile Photo

📉📉NEW SCALING LAW PHENOMENON 📉📉 We find that knowledge and reasoning exhibit different scaling behaviors! Super excited to finally tell you all about our paper on the compute optimal scaling of skills: arxiv.org/pdf/2503.10061 [1/n]

📉📉NEW SCALING LAW PHENOMENON 📉📉 

We find that knowledge and reasoning exhibit different scaling behaviors! 

Super excited to finally tell you all about our paper on the compute optimal scaling of skills: 
arxiv.org/pdf/2503.10061

[1/n]
Asher Trockman (@ashertrockman) 's Twitter Profile Photo

Are you a frontier lab investing untold sums in training? Are you trying to stay competitive? Are you finding that your competitors' models are ... thinking a bit too much like yours? Then antidistillation.com might be for you! Sam Altman Elon Musk

Are you a frontier lab investing untold sums in training? Are you trying to stay competitive? Are you finding that your competitors' models are ... thinking a bit too much like yours?

Then antidistillation.com might be for you! <a href="/sama/">Sam Altman</a> <a href="/elonmusk/">Elon Musk</a>
Calvin Luo (@calvinyluo) 's Twitter Profile Photo

Internet-scale datasets of videos and natural language are a rich training source! But can they be used to facilitate novel downstream robotic behaviors across embodiments and environments? Our new #ICLR2025 paper, Adapt2Act, shows how.

Marc Finzi (@m_finzi) 's Twitter Profile Photo

Why do larger language models generalize better? In our new ICLR paper, we derive an interpretable generalization bound showing that compute-optimal LLMs provably generalize better with scale! 📄arxiv.org/abs/2504.15208 1/7🧵

Yutong (Kelly) He (@electronickale) 's Twitter Profile Photo

✨ Love 4o-style image generation but prefer to use Midjourney? Tired of manual prompt crafting from inspo images? PRISM to the rescue! 🖼️→📝→🖼️ We automate black-box prompt engineering—no training, no embeddings, just accurate, readable prompts from your inspo images! 1/🧵

Runtian Zhai (@runtianzhai) 's Twitter Profile Photo

Why can foundation models transfer to so many downstream tasks? Will the scaling law end? Will pretraining end like Ilya Sutskever predicted? My PhD thesis builds the contexture theory to answer the above. Blog: runtianzhai.com/thesis Paper: arxiv.org/abs/2504.19792 🧵1/12

Stephen Bach (@stevebach) 's Twitter Profile Photo

🚀 Excited to share our new work on data generation for IR! We create synthetic multi-level ranking contexts for training dense retrievers. Now it’s easy to build custom retrieval datasets and move beyond the standard InfoNCE loss by learning from fine-grained relevance levels!🧵

🚀 Excited to share our new work on data generation for IR! We create synthetic multi-level ranking contexts for training dense retrievers. Now it’s easy to build custom retrieval datasets and move beyond the standard InfoNCE loss by learning from fine-grained relevance levels!🧵
Rattana Pukdee (@rpukdeee) 's Twitter Profile Photo

In our #AISTATS2025 paper, we ask: when it is possible to recover a consistent joint distribution from conditionals? We propose path consistency and autoregressive path consistency—necessary and easily verifiable conditions. See you at Poster session 3, Monday 5th May.

In our #AISTATS2025 paper, we ask: when it is possible to recover a consistent joint distribution from conditionals? We propose path consistency and autoregressive path consistency—necessary and easily verifiable conditions. 

See you at Poster session 3, Monday 5th May.
Xindi Wu (@cindy_x_wu) 's Twitter Profile Photo

Introducing COMPACT: COMPositional Atomic-to-complex Visual Capability Tuning, a data-efficient approach to improve multimodal models on complex visual tasks without scaling data volume. 📦 arxiv.org/abs/2504.21850 1/10

Introducing COMPACT: COMPositional Atomic-to-complex Visual Capability Tuning, a data-efficient approach to improve multimodal models on complex visual tasks without scaling data volume. 📦

arxiv.org/abs/2504.21850

1/10
Zhengyang Geng (@zhengyanggeng) 's Twitter Profile Photo

Excited to share our work with my amazing collaborators, Goodeat, Xingjian Bai, Zico Kolter, and Kaiming. In a word, we show an “identity learning” approach for generative modeling, by relating the instantaneous/average velocity in an identity. The resulting model,

Excited to share our work with my amazing collaborators, <a href="/Goodeat258/">Goodeat</a>, <a href="/SimulatedAnneal/">Xingjian Bai</a>, <a href="/zicokolter/">Zico Kolter</a>, and Kaiming.

In a word, we show an “identity learning” approach for generative modeling, by relating the instantaneous/average velocity in an identity. The resulting model,
Avi Schwarzschild (@a_v_i__s) 's Twitter Profile Photo

Big news! 🎉 I’m joining UNC-Chapel Hill as an Assistant Professor in Computer Science starting next year! Before that, I’ll be spending time OpenAI working on LLM privacy. UNC Computer Science UNC NLP

Big news! 🎉  I’m joining UNC-Chapel Hill as an Assistant Professor in Computer Science starting next year! Before that, I’ll be spending time <a href="/OpenAI/">OpenAI</a> working on LLM privacy.
<a href="/unccs/">UNC Computer Science</a> <a href="/uncnlp/">UNC NLP</a>
YixuanEvenXu (@yixuanevenxu) 's Twitter Profile Photo

✨ Did you know that NOT using all generated rollouts in GRPO can boost your reasoning LLM? Meet PODS! We down-sample rollouts and train on just a fraction, delivering notable gains over vanilla GRPO. (1/7)

✨ Did you know that NOT using all generated rollouts in GRPO can boost your reasoning LLM? Meet PODS! We down-sample rollouts and train on just a fraction, delivering notable gains over vanilla GRPO. (1/7)
Zhili Feng (@zhilifeng) 's Twitter Profile Photo

🥳🥳🥳I defended my PhD thesis today! Special thanks to my wonderful advisor Zico Kolter and committee members Russ Salakhutdinov Graham Neubig Lester Mackey! 🎉🎉🎉I am joining OpenAI as a researcher, super excited to keep working on frontier models and meet everyone in SF!

🥳🥳🥳I defended my PhD thesis today! Special thanks to my wonderful advisor <a href="/zicokolter/">Zico Kolter</a> and committee members <a href="/rsalakhu/">Russ Salakhutdinov</a> <a href="/gneubig/">Graham Neubig</a> <a href="/LesterMackey/">Lester Mackey</a>! 

🎉🎉🎉I am joining <a href="/OpenAI/">OpenAI</a> as a researcher, super excited to keep working on frontier models and meet everyone in SF!