Rabiraj Banerjee (@rabirajbandyop1) 's Twitter Profile
Rabiraj Banerjee

@rabirajbandyop1

NLP + CSS @gesis_org working on Hate Speech | ex Sr. Data Scientist @Coursera | Prev: MS in CS @UBuffalo
I aspire to use NLP for social media safety

ID: 1345873364561309697

calendar_today03-01-2021 23:23:22

403 Tweet

100 Followers

269 Following

Joshua Ong (@joshuaongg21) 's Twitter Profile Photo

We introduce PiCSAR (Probabilistic Confidence Selection And Ranking)💡: A simple training-free method for scoring samples based on probabilistic confidence, selecting a reasoning chain with the highest confidence from multiple sampled responses. ✏️PiCSAR is generalisable across

We introduce PiCSAR (Probabilistic Confidence Selection And Ranking)💡: A simple training-free method for scoring samples based on probabilistic confidence, selecting a reasoning chain with the highest confidence from multiple sampled responses.

✏️PiCSAR is generalisable across
Jason Weston (@jaseweston) 's Twitter Profile Photo

🌀Diversity Aware RL (DARLING)🌀 📝: arxiv.org/abs/2509.02534 - Jointly optimizes for quality & diversity using a learned partition function - Outperforms standard RL in quality AND diversity metrics, e.g. higher pass@1/p@k - Works for both non-verifiable & verifiable tasks 🧵1/5

🌀Diversity Aware RL (DARLING)🌀
📝: arxiv.org/abs/2509.02534
- Jointly optimizes for quality & diversity using a learned partition function
- Outperforms standard RL in quality AND diversity metrics, e.g. higher pass@1/p@k
- Works for both non-verifiable & verifiable tasks
🧵1/5
Kenan Tang (@kenantang) 's Twitter Profile Photo

Are LLMs really so prompt-sensitive? 🤔 🚨 Thrilled to share our EMNLP 2025 main conference paper! Prompt sensitivity has long been seen as a core weakness of LLMs—where tiny wording changes flip benchmark results. Our study finds: much of this effect stems from evaluation

Ahmad Beirami @ ICLR 2025 (@abeirami) 's Twitter Profile Photo

This is a great example of what good research looks like. You start with a real problem. You peel it layer by layer to find the root cause. You form a new hypothesis and keep digging. At the end, you have something insightful to share!

Arvindh Arun (@arvindh__a) 's Twitter Profile Photo

Why does horizon length grow exponentially as shown in the METR plot? Our new paper investigates this by isolating the execution capabilities of LLMs. Here's why you shouldn't be fooled by slowing progress on typical short-task benchmarks... 🧵

Why does horizon length grow exponentially as shown in the METR plot? 

Our new paper investigates this by isolating the execution capabilities of LLMs.

Here's why you shouldn't be fooled by slowing progress on typical short-task benchmarks... 🧵
Dylan Sam (@dylanjsam) 's Twitter Profile Photo

🚨Excited to introduce a major development in building safer language models: Safety Pretraining! Instead of post-hoc alignment, we take a step back and embed safety directly into pretraining. 🧵(1/n)

🚨Excited to introduce a major development in building safer language models: Safety Pretraining!

Instead of post-hoc alignment, we take a step back and embed safety directly into pretraining.

🧵(1/n)
Anshuman Mishra (@heyyanshuman) 's Twitter Profile Photo

You're in a ML Engineer interview at Anthropic, and the interviewer asks: "Your LLM inference is running out of GPU memory with long conversations. How do you fix this?" Here's how you answer:

Anand Bhattad (@anand_bhattad) 's Twitter Profile Photo

So You Want to Be an Academic? A couple of years into your PhD, but wondering: "Am I doing this right?" Most of the advice is aimed at graduating students. But there's far less for junior folks who are still finding their academic path. My candid takes: anandbhattad.github.io/blogs/jr_grads…

Surya Ganguli (@suryaganguli) 's Twitter Profile Photo

Teaching a new course Stanford University this quarter on explainable AI, motivated by neuroscience. I have curated a paper list 4 pages long (link in comment). What are your favorite papers on explainable AI/mechanistic interpretability that I am missing? Please comment or DM. thanks!

Teaching a new course <a href="/Stanford/">Stanford University</a> this quarter on explainable AI, motivated by neuroscience.  I have curated a paper list 4 pages long (link in comment).  What are your favorite papers on explainable AI/mechanistic interpretability that I am missing? Please comment or DM. thanks!
jack morris (@jxmnop) 's Twitter Profile Photo

best paper or blog i've read in a while, highly recommend! John is brilliant and his research sets an example for the rest of us. recently i too have been thinking deeply about how many bits might be learned via one step of RL or SFT.. if you're thinking about this too, lmk!

Ahmad Beirami @ ICLR 2025 (@abeirami) 's Twitter Profile Photo

Does each policy gradient update only contain O(1) bits of information about the model we are trying to train while each SFT update has O(T) where T is the number of tokens in the example? I think the answer is not a clean binary. Let’s examine this in a simple setup. Let’s

Yejin Choi (@yejinchoinka) 's Twitter Profile Photo

Instruction tuning has a hidden cost: ✅ Better at following instructions ❌ Narrower output distribution ❌ Worse in-context steerability We built 🌈 Spectrum Suite to investigate this and 🌈 Spectrum Tuning as an alternative post-training method —

Dylan Foster 🐢 (@canondetortugas) 's Twitter Profile Photo

The coverage principle: How pre-training enables post-training New preprint where we look at the mechanisms through which next-token prediction produces models that succeed at downstream tasks. The answer involves a metric we call the "coverage profile", not cross-entropy.

The coverage principle: How pre-training enables post-training

New preprint where we look at the mechanisms through which next-token prediction produces models that succeed at downstream tasks. 

The answer involves a metric we call the "coverage profile", not cross-entropy.
Jia-Bin Huang (@jbhuang0604) 's Twitter Profile Photo

Junior students who have just started doing research? Check out the (75 and counting) awesome tips! github.com/jbhuang0604/aw…

Junior students who have just started doing research? 

Check out the (75 and counting) awesome tips!

github.com/jbhuang0604/aw…
Mor Geva (@megamor2) 's Twitter Profile Photo

✨ New course materials: Interpretability of LLMs✨ This semester I'm teaching an active-learning grad course at Tel Aviv University on LLM interpretability, co-developed with my student Daniela Gottesman. We're releasing the materials as we go, so they can serve as a resource for anyone

elvis (@omarsar0) 's Twitter Profile Photo

This new research paper claims to complete million-step LLM tasks with zero errors. Huge for improving reliable long-chain AI reasoning. Worth checking out if you are an AI dev. Current LLMs degrade substantially when executing extended reasoning chains. Error rates compound

This new research paper claims to complete million-step LLM tasks with zero errors.

Huge for improving reliable long-chain AI reasoning.

Worth checking out if you are an AI dev.

Current LLMs degrade substantially when executing extended reasoning chains. Error rates compound