Seungwook Han (@seungwookh) 's Twitter Profile
Seungwook Han

@seungwookh

phd-ing @MIT_CSAIL, prev @MITIBMLab @columbia

ID: 876241258749894656

linkhttp://hanseungwook.github.io calendar_today18-06-2017 00:52:50

106 Tweet

319 Takipçi

488 Takip Edilen

Linlu Qiu (@linluqiu) 's Twitter Profile Photo

LLMs are increasingly used as agents that interact with users. To do so successfully, LLMs need to form beliefs and update them when new information becomes available. Do LLMs do so as expected from an optimal strategy? If not, can we get them to follow this strategy? 🧵

LLMs are increasingly used as agents that interact with users. To do so successfully, LLMs need to form beliefs and update them when new information becomes available. Do LLMs do so as expected from an optimal strategy? If not, can we get them to follow this strategy? 🧵
Seungwook Han (@seungwookh) 's Twitter Profile Photo

agreed that we'll eventually all have personalized models for each one of us -- just like how our feed, recommendations, and ads (regardless of whether it is good or bad) are being personalized to us

MIT NLP (@nlp_mit) 's Twitter Profile Photo

Hello everyone! We are quite a bit late to the twitter party, but welcome to the MIT NLP Group account! follow along for the latest research from our labs as we dive deep into language, learning, and logic 🤖📚🧠

Hello everyone! We are quite a bit late to the twitter party, but welcome to the MIT NLP Group account! follow along for the latest research from our labs as we dive deep into language, learning, and logic 🤖📚🧠
Pulkit Agrawal (@pulkitology) 's Twitter Profile Photo

Llama 4 (Meta) results are consistent with what we hypothesized will unleash the next generation of AI reasoning. A new paradigm for pre-training is around the corner arxiv.org/abs/2502.19402

Seungwook Han (@seungwookh) 's Twitter Profile Photo

In our recent paper, we hypothesized that SFT can limit downstream RL exploration and Llama 4 from Meta shows another convincing piece of evidence that this is true. Could this mean that next-token pretraining may be trapping us from training models that can truly reason? We

Seungwook Han (@seungwookh) 's Twitter Profile Photo

this is a great effort and we should be building towards a general platform like Universe from OpenAI to evaluate models on these games that inherently require different components of reasoning

Max Simchowitz (@max_simchowitz) 's Twitter Profile Photo

There’s a lot of awesome research about LLM reasoning right now. But how is  learning in the physical world 🤖different than in language 📚? In a new paper, show that imitation learning in continuous spaces can be exponentially harder than for discrete state spaces, even when

Ġabe Ġrand (@gabe_grand) 's Twitter Profile Photo

Tackling complex problems with LMs requires search/planning, but how should test-time compute be structured? Introducing Self-Steering, a new meta-reasoning framework where LMs coordinate their own inference procedures by writing code!

Mehul Damani @ ICLR (@mehuldamani2) 's Twitter Profile Photo

I am super excited to be presenting our work on adaptive inference -time compute at ICLR! Come chat with me on Thursday 4/24 at 3PM (Poster #219). I am also happy to chat about RL/reasoning/ RLHF/ inference scaling (DMs are open)!

Belinda Li @ ICLR 2025 (@belindazli) 's Twitter Profile Photo

I'll be presenting our work "Eliciting Human Preference with Language Models" at ICLR! Come catch my poster Thursday 4/24 at 10AM → iclr.cc/virtual/2025/p… Also DM me if you're interested in world models, interpretability, personalized interaction, or just general chatting!

I'll be presenting our work "Eliciting Human Preference with Language Models" at ICLR! Come catch my poster Thursday 4/24 at 10AM → iclr.cc/virtual/2025/p…

Also DM me if you're interested in world models, interpretability, personalized interaction, or just general chatting!
Shobhita Sundaram (@shobsund) 's Twitter Profile Photo

I'm at #ICLR2025 ! Excited to present our work on personalizing vision models with Julia Chae on Sat morning (poster #70). Please reach out if you want to chat about synthetic data (esp scaling, self-improvement, useful reasoning traces), rep learning, or anything else!

Hyojin Bahng (@hyojinbahng) 's Twitter Profile Photo

Image-text alignment is hard — especially as multimodal data gets more detailed. Most methods rely on human labels or proprietary feedback (e.g., GPT-4V). We introduce: 1. CycleReward: a new alignment metric focused on detailed captions, trained without human supervision. 2.

Image-text alignment is hard — especially as multimodal data gets more detailed. Most methods rely on human labels or proprietary feedback (e.g., GPT-4V).

We introduce:
1. CycleReward: a new alignment metric focused on detailed captions, trained without human supervision.
2.
Phillip Isola (@phillip_isola) 's Twitter Profile Photo

Our computer vision textbook is now available for free online here: visionbook.mit.edu We are working on adding some interactive components like search and (beta) integration with LLMs. Hope this is useful and feel free to submit Github issues to help us improve the text!