
Dylan X. Hou
@xinminghou
undergrad studying AI at Renmin Univ. of China, NLP researcher, intelligence explorer&trainer, interned@Tencent AI Lab. Carpe Diem🍀
ID: 1549288486435684353
https://dxhou.github.io/ 19-07-2022 07:02:19
486 Tweet
496 Followers
2,2K Following


New Research (w/ amazing Hangfeng He) "A Law of Next-Token Prediction in Large Language Models" LLMs rely on NTP, but their internal mechanisms seem chaotic. It's difficult to discern how each layer processes data for NTP. Surprisingly, we discover a physics-like law on NTP:



New research paper from Meta FAIR – Transfusion: Predict the Next Token and Diffuse Images with One Multi-Modal Model. Chunting Zhou, Lili Yu (ICLR2025) and team introduce this recipe for training a multi-modal model over discrete and continuous data. Transfusion combines next token


Should AI be aligned with human preferences, rewards, or utility functions? Excited to finally share a preprint that Micah Carroll Matija Hal Ashton & I have worked on for almost 2 years, arguing that AI alignment has to move beyond the preference-reward-utility nexus!









So excited to share that this was published in Nature Human Behaviour! 🥳 It's time to build AI thought partners that learn & think *with* people rather than *instead of* people. 🧠🤝🤖 We lay out what that means, why it matters, and how it can be done! nature.com/articles/s4156…

📣Proud to share Web2Code: a Large-scale Webpage-to-Code Dataset and Evaluation Framework for Multimodal LLMs was accepted to NeurIPS Conference 2024! About Web2Code: 📸 novel image + html dataset 📈webpage code gen benchmark 🧠CrystalChat-7B-Web2Code Blog: mbzuai-llm.github.io/webpage2code/


Really excited about my second Nature subjournal paper "Inferring Human Behavior from Language" nature.com/articles/s4156… Thanks a lot to Rada Mihalcea James W Pennebaker Laura Biester Ryan Boyd Verónica Pérez-Rosas Steve Wilson for this NLP+Psychology collaboration!





