Preetum Nakkiran (@preetumnakkiran) 's Twitter Profile
Preetum Nakkiran

@preetumnakkiran

ML research @ Apple

ID: 1069244826

linkhttp://preetum.nakkiran.org calendar_today07-01-2013 21:46:28

4,4K Tweet

11,11K Followers

2,2K Following

Eugene Vinitsky 🍒🦋 (@eugenevinitsky) 's Twitter Profile Photo

We now know RL agents can zero-shot crush driving benchmarks. Can we put them on a car and replace the planning stack? We're hiring a postdoc at NYU to find out! Email me if interested and please help us get the word out.

Miguel Angel Bautista (@itsbautistam) 's Twitter Profile Photo

We have an open position at Apple MLR to work scalable and efficient generative models that perform across diverse data domains—including images, 3D, video, graphs, etc. We care deeply about simplifying modeling pipelines, developing powerful and scalable training recipes.

Mehrdad Farajtabar (@mfarajtabar) 's Twitter Profile Photo

We have a full-time position for research scientist in our team at #Apple. The topic is understanding and improving #reasoning abilities of #LLMs. We're also interested in developing new and efficient architectures based on transformer for language modeling, again reasoning

Arwen Bradley (@arwenbradley) 's Twitter Profile Photo

I’m at #ICML2025 this week! Will be presenting Mechanism of Projective Composition of Diffusion Models tomorrow afternoon. Stop by poster E-3105 to see oil paintings of Preetum Nakkiran’s dog!

Shuangfei Zhai (@zhaisf) 's Twitter Profile Photo

There is a more general version of this question: why not scale up the parameters of the attention operation and make it more expressive? (you can do it as suggested below, or simply increase the dimension of QKV) The empirical answer is that it’s not nearly as effective as

Chhavi Yadav (@chhaviyadav_) 's Twitter Profile Photo

Had an amazing time NewInML @ ICML 2025 ICML Conference giving a talk on "What I Wish I knew before starting a PhD (but learnt the hard way)"! Loved the post-talk discussions and the heart warming messages :) Sharing slides since some people asked, link in the tweet below 👇

Had an amazing time <a href="/NewInML/">NewInML @ ICML 2025</a> <a href="/icmlconf/">ICML Conference</a> giving a talk on "What I Wish I knew before starting a PhD (but learnt the hard way)"!
Loved the post-talk discussions and the heart warming messages :)

Sharing slides since some people asked, link in the tweet below 👇
Christopher D. Long 🇺🇦 (@octonion) 's Twitter Profile Photo

Terence Tao has written a thread on Mathstodon about the damage being caused by the grant freezes at UCLA. mathstodon.xyz/@tao/114956840…

Dylan Foster 🐢 (@canondetortugas) 's Twitter Profile Photo

Announcing the first workshop on Foundations of Language Model Reasoning (FoRLM) at NeurIPS 2025! 📝Soliciting abstracts that advance foundational understanding of reasoning in language models, from theoretical analyses to rigorous empirical studies. 📆 Deadline: Sept 3, 2025

Announcing the first workshop on Foundations of Language Model Reasoning (FoRLM) at NeurIPS 2025!

📝Soliciting abstracts that advance foundational understanding of reasoning in language models, from theoretical analyses to rigorous empirical studies. 

📆 Deadline: Sept 3, 2025
Patrick McKenzie (@patio11) 's Twitter Profile Photo

One reason to not spend overly much time lawyering the meaning of words to minimize LLM’s capabilities is that you should not want to redefine thinking such that many humans have never thought.

Shreya Shankar (@sh_reya) 's Twitter Profile Photo

On my way to VLDB! 🇬🇧 I am on the job market this year, seeking tenure-track CS faculty positions. I will be giving a talk on DocETL and on a panel titled “Where Does Academic Database Research Go From Here?” I would love to meet folks; please reach out if you’re also attending!

Behnam Neyshabur (@bneyshabur) 's Twitter Profile Photo

OK, Sara Wiltberger and I are experimenting with a small, project-based mentorship program designed for the age of AI. We’re looking for resourceful self-starters—from early high school to early-career professionals—who want to prove their abilities through hard work. You don’t

Charles 🎉 Frye (@charles_irl) 's Twitter Profile Photo

The ICLR 2026 deadline is ten days away. But you just found a bug in your evals, so now you need to re-run all your ablations. That's hundreds of experiments, and you need them done ASAP. Modal's got you. Introducing our ICLR 2026 compute grant program.

The ICLR 2026 deadline is ten days away.

But you just found a bug in your evals, so now you need to re-run all your ablations. That's hundreds of experiments, and you need them done ASAP.

<a href="/modal/">Modal</a>'s got you.

Introducing our ICLR 2026 compute grant program.