Siddharth Suresh (@siddsuresh97) 's Twitter Profile
Siddharth Suresh

@siddsuresh97

PhD student @UWMadison | Applied Scientist Intern @AmazonScience AGI Foundations| Human-AI Alignment|Prev Intern @BrownCLPS

ID: 789065132122992640

linkhttps://scholar.google.com/citations?user=xsyrntwAAAAJ&hl=en calendar_today20-10-2016 11:26:02

133 Tweet

359 Takipçi

1,1K Takip Edilen

Computational Auditory Perception Group (@compaudition) 's Twitter Profile Photo

We are recruiting postdocs! Want to grow your own social networks to study creativity, cultural evolution & decision-making? We are hiring a funded postdoc at Cornell in collaboration with UC Davis, CUNY, & Princeton. Apply here: academicjobsonline.org/ajo/jobs/28959

Andrew Lampinen (@andrewlampinen) 's Twitter Profile Photo

Had fun talking at the Spurious Correlation & Shortcut Learning Workshop at ICLR! One example I brought up, which I think provides an uncommon perspective: a case where spurious shortcuts can improve generalization... even to out-of-distribution sets where the spurious feature doesn't generalize! Thread:

Dimitris Papailiopoulos (@dimitrispapail) 's Twitter Profile Photo

Do you want to do RL for coding and agentic workflows? Do you want to do science, and figure out when RL kicks in? What is the right algorithm (it's not GRPO)? how much reasoning you need in your base (you def need some! but is it a lot or A LOT)? Do you want to figure out how

Andrew Lampinen (@andrewlampinen) 's Twitter Profile Photo

How do language models generalize from information they learn in-context vs. via finetuning? We show that in-context learning can generalize more flexibly, illustrating key differences in the inductive biases of these modes of learning — and ways to improve finetuning. Thread: 1/

How do language models generalize from information they learn in-context vs. via finetuning? We show that in-context learning can generalize more flexibly, illustrating key differences in the inductive biases of these modes of learning — and ways to improve finetuning. Thread: 1/
Qihong Lu | 吕其鸿 (@qihong_lu) 's Twitter Profile Photo

I’m thrilled to announce that I will start as a presidential assistant professor in Neuroscience at the City U of Hong Kong in Jan 2026! I have RA, PhD, and postdoc positions available! Come work with me on neural network models/experiments on human memory! RT appreciated! (1/5)

Fenil Doshi (@fenildoshi009) 's Twitter Profile Photo

🧵 What if two images have the same local parts but represent different global shapes purely through part arrangement? Humans can spot the difference instantly! The question is can vision models do the same? 1/15

Jifan Zhang (@jifan_zhang) 's Twitter Profile Photo

Releasing HumorBench today. Grok 4 is🥇 on this uncontaminated, non-STEM humor reasoning benchmark. 🫡🫡xAI Here are couple things I find surprising👇 1. this benchmark yields an almost perfect rank correlation with ARC-AGI. Yet the task of reasoning about New Yorker style

Thomas Fel (@napoolar) 's Twitter Profile Photo

🧠 Submit to CogInterp @ NeurIPS 2025! Bridging AI & cognitive science to understand how models think, reason & represent. CFP + details 👉 coginterp.github.io/neurips2025/

Nicholas Roberts (@nick11roberts) 's Twitter Profile Photo

🎉 Excited to share that our paper "Pretrained Hybrids with MAD Skills" was accepted to Conference on Language Modeling 2025! We introduce Manticore - a framework for automatically creating hybrid LMs from pretrained models without training from scratch. 🧵[1/n]

lalit (@stochasticlalit) 's Twitter Profile Photo

It was amazing to be part of this effort. Huge shout out to the team, and all the incredible pre-training and post-training efforts that ensure Gemini is the leading frontier model! deepmind.google/discover/blog/…

Cameron R. Wolfe, Ph.D. (@cwolferesearch) 's Twitter Profile Photo

Direct Preference Optimization (DPO) is simple to implement but complex to understand, which creates misconceptions about how it actually works… LLM Training Stages: LLMs are typically trained in four stages: 1. Pretraining 2. Supervised Finetuning (SFT) 3. Reinforcement

Direct Preference Optimization (DPO) is simple to implement but complex to understand, which creates misconceptions about how it actually works…

LLM Training Stages: LLMs are typically trained in four stages:

1. Pretraining
2. Supervised Finetuning (SFT)
3. Reinforcement
Apurva Ratan Murty (@apurvaratan) 's Twitter Profile Photo

Excited for this workshop at #CCN2025! Come listen to me talk about TopoNets: Topographic models across vision, language and audition. Look forward to seeing old friends and making new ones!