Deep-Learning Times 💙 (@s_chatterjee66) 's Twitter Profile
Deep-Learning Times 💙

@s_chatterjee66

Incoming Researcher @ KrutrimLLM || CS @ Indian Statistical Institute || A researcher in deep learning. Looking for collaborations
Let's connect !!

ID: 1916431237960413184

linkhttps://www.linkedin.com/in/sandeep-chatterjee-918290143/ calendar_today27-04-2025 09:57:15

89 Tweet

16 Followers

420 Following

Niloofar (on faculty job market!) (@niloofar_mire) 's Twitter Profile Photo

📣Thrilled to announce I’ll join Carnegie Mellon University (CMU Engineering & Public Policy & Language Technologies Institute | @CarnegieMellon) as an Assistant Professor starting Fall 2026! Until then, I’ll be a Research Scientist at AI at Meta FAIR in SF, working with Kamalika Chaudhuri’s amazing team on privacy, security, and reasoning in LLMs!

📣Thrilled to announce I’ll join Carnegie Mellon University (<a href="/CMU_EPP/">CMU Engineering & Public Policy</a> &amp; <a href="/LTIatCMU/">Language Technologies Institute | @CarnegieMellon</a>) as an Assistant Professor starting Fall 2026!

Until then, I’ll be a Research Scientist at <a href="/AIatMeta/">AI at Meta</a> FAIR in SF, working with <a href="/kamalikac/">Kamalika Chaudhuri</a>’s amazing team on privacy, security, and reasoning in LLMs!
Yujin Kim (@yujin301300) 's Twitter Profile Photo

Introducing our new work: 🚀Mixture-of-Recursions! 🪄We propose a novel framework that dynamically allocates recursion depth per token. 🪄MoR is an efficient architecture with fewer params, reduced KV cache memory, and 2× greater throughput— maintaining comparable performance!

Introducing our new work: 🚀Mixture-of-Recursions!

🪄We propose a novel framework that dynamically allocates recursion depth per token.

🪄MoR is an efficient architecture with fewer params, reduced KV cache memory, and 2× greater throughput— maintaining comparable performance!
Edoardo Debenedetti (@edoardo_debe) 's Twitter Profile Photo

Excited to start as a Research Scientist Intern at Meta, in the GenAI Red Team, where I will keep working on AI agents security. I'll be based in the Bay Area, so reach out if you're around and wanna chat about AI security!

Excited to start as a Research Scientist Intern at Meta, in the GenAI Red Team, where I will keep working on AI agents security. I'll be based in the Bay Area, so reach out if you're around and wanna chat about AI security!
Nate Chen (@chengua46724992) 's Twitter Profile Photo

Why do FFNs use ReLU instead of more precise ones like Exp? "We propose the following hypothesis: A kernel with lower retrieval precision encourages a more polysemantic key–value memory: multiple unrelated facts can be stored under the same key space" Great and inspiring read!

Why do FFNs use ReLU instead of more precise ones like Exp?

"We propose the following hypothesis: A kernel with lower retrieval precision encourages a more polysemantic key–value memory: multiple unrelated facts can be stored under the same key space"

Great and inspiring read!
Yam Peleg (@yampeleg) 's Twitter Profile Photo

Wild paper They prove (!!) a transformer block (Attn + MLP) running on prompt Outputs the same logits with no prompt If MLP weights updated by vector: W′ = W + ΔW Calc from attn latent: ΔW = (W·Δa) × (A(x)ᵀ / ‖A(x)‖²) Given prompt: Δa = A(C, x) − A(x) Fucking fine tuning.

Wild paper

They prove (!!) a transformer block (Attn + MLP) running on prompt

Outputs the same logits with no prompt

If MLP weights updated by vector:
W′ = W + ΔW

Calc from attn latent:
ΔW = (W·Δa) × (A(x)ᵀ / ‖A(x)‖²)

Given prompt:
Δa = A(C, x) − A(x)

Fucking fine tuning.
alphaXiv (@askalphaxiv) 's Twitter Profile Photo

In-context learning is just gradient descent without explicit training! This paper "Learning without training: The implicit dynamics of in-context learning" shows that ICL can be mathematically interpreted as an implicit low-rank weight update during inference.

In-context learning is just gradient descent without explicit training!

This paper "Learning without training: The implicit dynamics of in-context learning" shows that ICL can be mathematically interpreted as an implicit low-rank weight update during inference.
ℏεsam (@hesamation) 's Twitter Profile Photo

Fuck ML tutorials. This is a collection of 300 ML system design case studies in real world, from Stripe, Spotify, Netflix, Meta, etc. Perfect for interviews and to learn how it’s done in the battlefield. Wish there was a similar thing for agents!

Fuck ML tutorials. 

This is a collection of 300 ML system design case studies in real world, from Stripe, Spotify, Netflix, Meta, etc.

Perfect for interviews and to learn how it’s done in the battlefield. Wish there was a similar thing for agents!
Jackson Atkins (@jacksonatkinsx) 's Twitter Profile Photo

LLMs can now self-optimize. A new method allows an AI to rewrite its own prompts to achieve up to 35x greater efficiency, outperforming both Reinforcement Learning and Fine-Tuning for complex reasoning. UC Berkeley, Stanford, and Databricks introduce a new method called GEPA

LLMs can now self-optimize. 

A new method allows an AI to rewrite its own prompts to achieve up to 35x greater efficiency, outperforming both Reinforcement Learning and Fine-Tuning for complex reasoning.

UC Berkeley, Stanford, and Databricks introduce a new method called GEPA
Rosinality (@rosinality) 's Twitter Profile Photo

Geometric-Mean Policy Optimization Using geometric mean for the importance ratio, similar to GSPO (arxiv.org/abs/2507.18071).

Geometric-Mean Policy Optimization

Using geometric mean for the importance ratio, similar to GSPO (arxiv.org/abs/2507.18071).
Sumanth (@sumanth_077) 's Twitter Profile Photo

Build a Large Language Model from scratch! This repository contains the code for developing, pretraining, and finetuning a GPT-like large language model. 100% Free & Open Source

Build a Large Language Model from scratch!

This repository contains the code for developing, pretraining, and finetuning a GPT-like large language model.

100% Free &amp; Open Source
Dr. PM Dhakate (@paragenetics) 's Twitter Profile Photo

An amazing video, our national animal and bird, together in one frame! A perfect symbol of India's vibrant spirit. Wishing everyone a Happy Independence Day. आप सभी को स्वतंत्रता दिवस की हार्दिक बधाई एवं शुभकामनाएं, जय हिंद। 🇮🇳 VC: Rakesh Bhatt #IndependenceDay #JaiHind