Abhimanyu Hans (@ahans30) 's Twitter Profile
Abhimanyu Hans

@ahans30

CS PhD student advised by @tomgoldsteincs at @umdcs

ID: 328935348

linkhttp://ahans30.github.io calendar_today04-07-2011 08:51:29

188 Tweet

220 Takipçi

1,1K Takip Edilen

Tanishq Mathew Abraham, Ph.D. (@iscienceluvr) 's Twitter Profile Photo

Refusal Tokens: A Simple Way to Calibrate Refusals in Large Language Models "We introduce a simple strategy that makes refusal behavior controllable at test-time without retraining: the refusal token. During alignment, we prepend a special [refuse] token to responses that

Refusal Tokens: A Simple Way to Calibrate Refusals in Large Language Models

"We introduce a simple strategy that makes refusal behavior controllable at test-time without retraining: the refusal token. During alignment, we prepend a special [refuse] token to responses that
Tom Goldstein (@tomgoldsteincs) 's Twitter Profile Photo

For those a #NeurIPS2024 I'm giving a talk at 2pm today (tues) on how to build a successful research project from ideation to paper submission. Other speakers include researchers Maarten Sap (he/him) and Kaiming He, plus Nicole Bannon from the hiring firm Rora (Formerly Ralph) Inc. See you at NewInML @ ICML 2025

Avi Schwarzschild (@a_v_i__s) 's Twitter Profile Photo

I’m on the faculty job market this year and I’m on my way to NeurIPS! Let’s talk about AI safety and reasoning capability.

Neel Jain (@neeljain1717) 's Twitter Profile Photo

Excited to present “Be Like a Goldfish: Don’t Memorize!” led by Abhimanyu Hans at NeurIPS, East Building #4709. In this work, we propose a simple yet effective loss function to reduce verbatim memorization. Stop by the poster and grab some goldfish crackers while you’re at it!

Excited to present “Be Like a Goldfish: Don’t Memorize!” led by <a href="/ahans30/">Abhimanyu Hans</a> at NeurIPS, East Building #4709. In this work, we propose a simple yet effective loss function to reduce verbatim memorization.

Stop by the poster and grab some goldfish crackers while you’re at it!
Sean McLeish (@seanmcleish) 's Twitter Profile Photo

Why is addition hard for next token predictors? Come hear about our fix for this at #NeurIPS! We’re presenting Abacus Embeddings 🧮 tomorrow (Friday) from 11-2 in the East Exhibit hall at poster 2907. Drop by to see how we can improve your language model.

Vatsal Baherwani (@vatsalbaherwani) 's Twitter Profile Photo

MoEs are computationally efficient, but their sparse gradients cause training instability. Our dense gradient approximation improves stability and performance w/ minimal overhead. Come see our #NeurIPS2024 poster today at the ENLSP workshop and tomorrow at the OPT-ML workshop!

MoEs are computationally efficient, but their sparse gradients cause training instability. Our dense gradient approximation improves stability and performance w/ minimal overhead.

Come see our #NeurIPS2024 poster today at the ENLSP workshop and tomorrow at the OPT-ML workshop!
Neel Jain (@neeljain1717) 's Twitter Profile Photo

Will be presenting Refusal Tokens at the Safe GenAI #NeurIPS2024 workshop tomorrow in the East Exhibition Hall A Come drop by and say hi!

Vikash Sehwag (@vsehwag_) 's Twitter Profile Photo

Following fully open-source philosophy, we’ve released the official training code, data code, and model ckpts for our micro-budget training of diffusion models from scratch (MicroDiTs). Now anyone can train a Stable Diffusion v1/v2-quality model from scratch in just 2.5 days

Following fully open-source philosophy, we’ve released the official training code, data code, and model ckpts for our micro-budget training of diffusion models from scratch (MicroDiTs).

Now anyone can train a Stable Diffusion v1/v2-quality model from scratch in just 2.5 days
Micah Goldblum (@micahgoldblum) 's Twitter Profile Photo

🚨📢 Excited to announce the ICLR 2025 Workshop on Building Trust in LLMs and LLM Applications! 📢🚨 Submit all your papers, and we’ll see you in Singapore! There will be paper awards, and we have a stacked lineup of speakers and panelists.

Tom Goldstein (@tomgoldsteincs) 's Twitter Profile Photo

Let’s sanity check DeepSeek’s claim to train on 2048 GPUs for under 2 months, for a cost of $5.6M. It sort of checks out and sort of doesn't. The v3 model is an MoE with 37B (out of 671B) active parameters. Let's compare to the cost of a 34B dense model. 🧵 (1/4)

Ashwinee Panda (@pandaashwinee) 's Twitter Profile Photo

1 week to submit to our Sparsity workshop at ICLR 2026! That means SAEs, Sparse models, KV cache compression, quantization, pruning; we want to bring together folks from different sub areas to share ideas on how to make LLMs smaller, faster, and better!

Abhimanyu Hans (@ahans30) 's Twitter Profile Photo

the real problem is that current retrievers are not instruction-friendly that's why you google keywords likely to be found in the answers/docs than try to "prompt" your query or you know...just ask your LLM

Tom Goldstein (@tomgoldsteincs) 's Twitter Profile Photo

New open source reasoning model! Huginn-3.5B reasons implicitly in latent space 🧠 Unlike O1 and R1, latent reasoning doesn’t need special chain-of-thought training data, and doesn't produce extra CoT tokens at test time. We trained on 800B tokens 👇

New open source reasoning model!

Huginn-3.5B reasons implicitly in latent space 🧠

Unlike O1 and R1, latent reasoning doesn’t need special chain-of-thought training data, and doesn't produce extra CoT tokens at test time.

We trained on 800B tokens 👇
Max Spero (@max_spero_) 's Twitter Profile Photo

Peer reviewers are swamped. What if we gave them extra work by injecting AI-generated papers into the system? Oh, and in the name of ethics we can let them know ahead of time that some papers are AI slop. Just to keep everyone paranoid.

Maharshi Gor (@maharshigor) 's Twitter Profile Photo

Can you spot when AI bluffs?🤖 Can you outguess AI—or work with one to dominate trivia?🏁 🏆 We are hosting the first Human–AI coop trivia (Quizzing) competition. 🎲Play, 🛠️build, or ✍🏼write questions... ..and win prizes 🎁. 🥳 It’s fun, free, and happening this June 🧠🤖👇

Can you spot when AI bluffs?🤖 Can you outguess AI—or work with one to dominate trivia?🏁

🏆 We are hosting the first Human–AI coop trivia (Quizzing) competition.

🎲Play, 🛠️build, or ✍🏼write questions... 
..and win prizes 🎁.

🥳 It’s fun, free, and happening this June 🧠🤖👇
Avi Schwarzschild (@a_v_i__s) 's Twitter Profile Photo

Big news! 🎉 I’m joining UNC-Chapel Hill as an Assistant Professor in Computer Science starting next year! Before that, I’ll be spending time OpenAI working on LLM privacy. UNC Computer Science UNC NLP

Big news! 🎉  I’m joining UNC-Chapel Hill as an Assistant Professor in Computer Science starting next year! Before that, I’ll be spending time <a href="/OpenAI/">OpenAI</a> working on LLM privacy.
<a href="/unccs/">UNC Computer Science</a> <a href="/uncnlp/">UNC NLP</a>
Furong Huang (@furongh) 's Twitter Profile Photo

There’s been heated debate lately: Can generative AI truly self-improve? ✅Some say yes, pointing to models learning like curious humans. ❌Others say no, invoking the first law of thermodynamics: You can’t get something from nothing. No new info, no gain. 🧠 But what if the

Abhimanyu Hans (@ahans30) 's Twitter Profile Photo

zoom bombing is lame guys, especially in 2025, especially in someone's PhD proposal talk totally unrelated but guess who's a PhD candidate now 👀

Thao Nguyen (@thao_nguyen26) 's Twitter Profile Photo

We released 44B synthetic tokens from our CoT-guided rewriting, offering higher quality pretraining data than the average human-written web texts📈 🤗Data: huggingface.co/datasets/faceb… 📜Paper: arxiv.org/abs/2506.04689 (accepted at #COLM2025) Excited to see what the community builds!