Samyak (@sams_jain) 's Twitter Profile
Samyak

@sams_jain

Research Fellow @MSFTResearch, Previous @_FiveAI, @kasl_ai, @val_iisc, CS @IITBHU_Varanasi.

Interested in foundations of AI and AI Safety.

ID: 1355198417673150469

linkhttps://samyakjain0112.github.io/ calendar_today29-01-2021 16:57:47

102 Tweet

252 Takipçi

803 Takip Edilen

Abhishek Panigrahi (@abhishek_034) 's Twitter Profile Photo

Progressive distillation, where a student model learns from multiple checkpoints of the teacher, has been shown to improve the student–but why? We show it induces an implicit curriculum that accelerates training. Work w Bingbin Liu, Sadhika Malladi, Andrej Risteski, Surbhi Goel

Progressive distillation, where a student model learns from multiple checkpoints of the teacher, has been shown to improve the student–but why? We show it induces an implicit curriculum that  accelerates training. 

Work w <a href="/BingbinL/">Bingbin Liu</a>, <a href="/SadhikaMalladi/">Sadhika Malladi</a>, <a href="/risteski_a/">Andrej Risteski</a>, <a href="/SurbhiGoel_/">Surbhi Goel</a>
Ekdeep Singh Lubana (@ekdeepl) 's Twitter Profile Photo

Paper alert—accepted as a NeurIPS *Spotlight*!🧵👇 We build on our past work relating emergence to task compositionality and analyze the *learning dynamics* of such tasks: we find there exist latent interventions that can elicit them much before input prompting works! 🤯

Usman Anwar (@usmananwar391) 's Twitter Profile Photo

Transformers are REALLY good at in-context learning (ICL); but do they learn ‘adversarially robust’ ICL algorithms? We study this and much more in our new paper! 🧵

Transformers are REALLY good at in-context learning (ICL); but do they learn ‘adversarially robust’ ICL algorithms? We study this and much more in our new paper! 🧵
P Shravan Nayak (@pshravannayak) 's Twitter Profile Photo

Excited to be at #EMNLP2024! 🎉 Join my talk on CulturalVQA, a benchmark testing Vision Language Models’ grasp of cultural understanding. Let’s see if VLMs truly capture global perspectives—chat after! 🗓️ Nov 12 (Tue), 4:15-4:30 PM 📍 Flagler Paper: arxiv.org/abs/2407.10920

Samyak (@sams_jain) 's Twitter Profile Photo

I'll be at NeurIPS Conference, sharing my work on understanding safety fine-tuning and jailbreaks. Visit our poster on Wed 11, Session-2 (#3306) Also super excited to discuss about my current work at Microsoft Research on understanding lottery ticket hypothesis. Please reach out to chat!

Satwik Bhattamishra (@satwik1729) 's Twitter Profile Photo

Excited to head to NeurIPS Conference today! I'll be presenting our work on the representational capabilities of Transformers and RNNs/SSMs. If you're interested in meeting up to discuss research or chat, feel free to reach out via DM or email!

Excited to head to <a href="/NeurIPSConf/">NeurIPS Conference</a> today! I'll be presenting our work on the representational capabilities of Transformers and RNNs/SSMs. If you're interested in meeting up to discuss research or chat, feel free to reach out via DM or email!
Samyak (@sams_jain) 's Twitter Profile Photo

Very interesting work with several real world applications including designing natural jailbreaks and augmenting safety fine-tuning datasets.

Ekdeep Singh Lubana (@ekdeepl) 's Twitter Profile Photo

Paper alert––*Awarded best paper* at NeurIPS workshop on Foundation Model Interventions! 🧵👇 We analyze the (in)abilities of SAEs by relating them to the field of disentangled rep. learning, where limitations of AE based interpretability protocols have been well established!🤯

iKDD (@ikdd_news) 's Twitter Profile Photo

IKDD congratulates Sravanti Addepalli IISc Bangalore  for enhancing the robustness of Deep Neural Networks against adversarial attacks and distribution shifts while addressing practical deployment challenges. Preethi Jyothi Manish Gupta Amith Singhee

IKDD congratulates Sravanti Addepalli <a href="/iiscbangalore/">IISc Bangalore</a>  for enhancing the robustness of Deep Neural Networks against adversarial attacks and distribution shifts while addressing practical deployment challenges.

<a href="/PreethiJyothi1/">Preethi Jyothi</a> <a href="/ManishGuptaMG1/">Manish Gupta</a> <a href="/asinghee1/">Amith Singhee</a>
Andrew Lee (@a_jy_l) 's Twitter Profile Photo

New paper 🥳🚨 Interested in inference-time scaling? In-context Learning? Mech Interp? LMs can solve novel in-context tasks, with sufficient examples (longer contexts). Why? Bcus they dynamically form *in-context representations*! 1/N

Tim Rocktäschel (@_rockt) 's Twitter Profile Photo

Proud to announce that Dr Robert Kirk defended his PhD thesis titled "Understanding and Evaluating Generalisation for Superhuman AI Systems" last week 🥳. Massive thanks to Roger Grosse and Sebastian Riedel (@[email protected]) for examining! As is customary, Rob received a personal mortarboard from

Proud to announce that Dr <a href="/_robertkirk/">Robert Kirk</a> defended his PhD thesis titled "Understanding and Evaluating
Generalisation for Superhuman AI Systems" last week 🥳. Massive thanks to <a href="/RogerGrosse/">Roger Grosse</a> and <a href="/riedelcastro/">Sebastian Riedel (@riedelcastro@sigmoid.social)</a> for examining! As is customary, Rob received a personal mortarboard from
Samyak (@sams_jain) 's Twitter Profile Photo

This work from my friend Pranav Nair seems very interesting! I wonder how this relates to recent work on scaling laws for precision: arxiv.org/pdf/2411.04330. Does this follow a similar trend still? I guess for 2 bit quantization it breaks the expected trend.

Ekdeep Singh Lubana (@ekdeepl) 's Twitter Profile Photo

New paper–Accepted at #ICLR2025 and also my last PhD paper! 🧑‍🎓🧵👇 We propose a novel model of how emergent learning curves show up in neural nets’ training by making a connection to the theory of graph percolation!

New paper–Accepted at #ICLR2025 and also my last PhD paper! 🧑‍🎓🧵👇

We propose a novel model of how emergent learning curves show up in neural nets’ training by making a connection to the theory of graph percolation!
Sachin Yadav (@sachinyv) 's Twitter Profile Photo

✨New Paper: Presenting Interleaved Gibbs Diffusion (IGD), a novel generative framework for mixed continuous-discrete data, focusing on constrained generation. From 3-SAT and molecule design to layout generation, IGD advances diffusion models by capturing complex inter-variable

Andrew Lee (@a_jy_l) 's Twitter Profile Photo

🚨New preprint! How do reasoning models verify their own CoT? We reverse-engineer LMs and find critical components and subspaces needed for self-verification! 1/n

🚨New preprint!

How do reasoning models verify their own CoT?
We reverse-engineer LMs and find critical components and subspaces needed for self-verification!

1/n
Ekdeep Singh Lubana (@ekdeepl) 's Twitter Profile Photo

🚨 New paper alert! Linear representation hypothesis (LRH) argues concepts are encoded as **sparse sum of orthogonal directions**, motivating interpretability tools like SAEs. But what if some concepts don’t fit that mold? Would SAEs capture them? 🤔 1/11