Devansh Jain (@devanshrjain) 's Twitter Profile
Devansh Jain

@devanshrjain

model routers @ ยฌโ—‡ | ai safety @LTIatCMU | ex cs @bitspilaniindia

ID: 3301689482

calendar_today30-07-2015 16:08:00

89 Tweet

166 Followers

816 Following

Harshita Diddee (@ihsrahedid) 's Twitter Profile Photo

Ever wondered which instruction selection strategy to choose for your custom setup? The answer might just be random sampling! In our recent #NAACL Findings paper, we show that popular strategies do not *consistently* beat random selection! Paper: shorturl.at/77ECJ 1/6

Ever wondered which instruction selection strategy to choose for your custom setup? The answer might just be random sampling! In our recent #NAACL Findings paper, we show that popular strategies do not *consistently* beat random selection!
Paper: shorturl.at/77ECJ 1/6
Dagster (@dagster) 's Twitter Profile Photo

Deploying LLM applications isn't just about the initial setup; it's about continuously managing the technology's rapid evolution. As new models emerge, organizations face the challenge of minimizing technical debt and optimizing model utilization. Check out the full Deep Dive

Tomas Hernando Kofman (@tomas_hk) 's Twitter Profile Photo

We're hiring engineers and researchers to build the future of multi-model AI infrastructure. We're a small, technically elite team backed by Jeff Dean, Julien Chaumond, Ion Stoica, + more. And we guarantee a $50K investment in your next startup for every year you work with us.

MatthewBerman (@matthewberman) 's Twitter Profile Photo

Apply to NotDiamond, they are hiring the best engineers and researchers. I met Tomas and his vision and intelligence blew me away, so I personally invested! Plus, they will invest $50k into the next company you start for every year you work there.

Akhila Yerukola (@akhila_yerukola) 's Twitter Profile Photo

Did you know? Gestures to express universal conceptsโ€”like wishing for luckโ€”vary WIDELY across cultures? ๐Ÿคžmeans luck in US but deeply offensive in Vietnam ๐Ÿšจ ๐Ÿ“ฃWe introduce MC-SIGNS, a test bed to evaluate how LLMs/VLMs/T2I handle such nonverbal cues ๐Ÿ“œ: arxiv.org/abs/2502.17710

Did you know? Gestures to express universal conceptsโ€”like wishing for luckโ€”vary WIDELY across cultures?
๐Ÿคžmeans luck in US but deeply offensive in Vietnam ๐Ÿšจ

๐Ÿ“ฃWe introduce MC-SIGNS, a test bed to evaluate how LLMs/VLMs/T2I handle such nonverbal cues
๐Ÿ“œ: arxiv.org/abs/2502.17710
Neel Bhandari (@neelbhandari9) 's Twitter Profile Photo

1/๐Ÿšจ ๐—ก๐—ฒ๐˜„ ๐—ฝ๐—ฎ๐—ฝ๐—ฒ๐—ฟ ๐—ฎ๐—น๐—ฒ๐—ฟ๐˜ ๐Ÿšจ RAG systems excel on academic benchmarks - but are they robust to variations in linguistic style? We find RAG systems are brittle. Small shifts in phrasing trigger cascading errors, driven by the complexity of the RAG pipeline ๐Ÿงต

1/๐Ÿšจ ๐—ก๐—ฒ๐˜„ ๐—ฝ๐—ฎ๐—ฝ๐—ฒ๐—ฟ ๐—ฎ๐—น๐—ฒ๐—ฟ๐˜ ๐Ÿšจ
RAG systems excel on academic benchmarks - but are they robust to variations in linguistic style?

We find RAG systems are brittle. Small shifts in phrasing trigger cascading errors, driven by the complexity of the RAG pipeline ๐Ÿงต
MatthewBerman (@matthewberman) 's Twitter Profile Photo

Not Diamond is building an incredibly important infrastructure layer for AI: model routing. Today, they make it easier to write prompts once and use them across different LLMs with Prompt Adaptation. I got a preview a few weeks ago (Iโ€™m a small investor) and was very

Yong Zheng-Xin (Yong) (@yong_zhengxin) 's Twitter Profile Photo

๐Ÿงต Multilingual safety training/eval is now standard practice, but a critical question remains: Is multilingual safety actually solved? Our new survey with Cohere Labs answers this and dives deep into: - Language gap in safety research - Future priority areas Thread ๐Ÿ‘‡

๐Ÿงต Multilingual safety training/eval is now standard practice, but a critical question remains: Is multilingual safety actually solved?

Our new survey with <a href="/Cohere_Labs/">Cohere Labs</a> answers this and dives deep into:
- Language gap in safety research
- Future priority areas

Thread ๐Ÿ‘‡
Akhila Yerukola (@akhila_yerukola) 's Twitter Profile Photo

Thanks Language Technologies Institute | @CarnegieMellon and CMU School of Computer Science for featuring our work!!โœจ๐Ÿ’ซ Our paper on culturally offensive nonverbal gestures is accepted to #ACL2025 main! Detailed thread๐Ÿงต: x.com/akhila_yerukolโ€ฆ Preprint๐Ÿ“œ: arxiv.org/abs/2502.17710 Work done with Saadia Gabriel Violet Peng Maarten Sap (he/him)

Sanidhya Vijayvargiya (@sanidhya903) 's Twitter Profile Photo

1/ AI agents are increasingly being deployed for real-world tasks, but how safe are they in high-stakes settings? ๐Ÿšจ NEW: OpenAgentSafety - A comprehensive framework for evaluating AI agent safety in realistic scenarios across eight critical risk categories. ๐Ÿงต

1/ AI agents are increasingly being deployed for real-world tasks, but how safe are they in high-stakes settings?
๐Ÿšจ NEW: OpenAgentSafety - A comprehensive framework for evaluating AI agent safety in realistic scenarios across eight critical risk categories.
๐Ÿงต
Andy Liu (@uilydna) 's Twitter Profile Photo

๐ŸšจNew Paper: LLM developers aim to align models with values like helpfulness or harmlessness. But when these conflict, which values do models choose to support? We introduce ConflictScope, a fully-automated evaluation pipeline that reveals how models rank values under conflict.

๐ŸšจNew Paper: LLM developers aim to align models with values like helpfulness or harmlessness. But when these conflict, which values do models choose to support? We introduce ConflictScope, a fully-automated evaluation pipeline that reveals how models rank values under conflict.
Maarten Sap (he/him) (@maartensap) 's Twitter Profile Photo

Day 3 (Thu Oct 9), 11:00amโ€“1:00pm, Poster Session 5 Poster #13: PolyGuard: A Multilingual Safety Moderation Tool for 17 Languages โ€” led by Priyanshu Kumar, Devansh Jain Poster #74: Fluid Language Model Benchmarking โ€” led by Valentin Hofmann

Liwei Jiang (@liweijianglw) 's Twitter Profile Photo

(Thu Oct 9, 11:00amโ€“1:00pm) Poster Session 5 ๐๐จ๐ฌ๐ญ๐ž๐ซ #๐Ÿ๐Ÿ‘: PolyGuard: A Multilingual Safety Moderation Tool for 17 Languages; w/ amazing Priyanshu Kumar, Devansh Jain PolyGuard is among the SOTA multilingual safety moderation tool + we release comprehensive multilingual

(Thu Oct 9, 11:00amโ€“1:00pm) Poster Session 5

๐๐จ๐ฌ๐ญ๐ž๐ซ #๐Ÿ๐Ÿ‘: PolyGuard: A Multilingual Safety Moderation Tool for 17 Languages; w/ amazing <a href="/kpriyanshu256/">Priyanshu Kumar</a>, <a href="/devanshrjain/">Devansh Jain</a> 

PolyGuard is among the SOTA multilingual safety moderation tool + we release comprehensive multilingual
Rootly (@rootlyhq) 's Twitter Profile Photo

While Sonnet-4.5 remains a popular choice among developers, our benchmarks show it underperforms GPT-5 on SRE-related tasks when both are run with default parameters. However, using the Not Diamond prompt adaptation platform, Sonnet-4.5 achieved up to a 2x performance

Kshitish Ghate (@ghatekshitish) 's Twitter Profile Photo

๐ŸšจNew paper: Reward Models (RMs) are used to align LLMs, but can they be steered toward user-specific value/style preferences? With EVALUESTEER, we find even the best RMs we tested exhibit their own value/style biases, and are unable to align with a user >25% of the time. ๐Ÿงต

๐ŸšจNew paper: Reward Models (RMs) are used to align LLMs, but can they be steered toward user-specific value/style preferences? 
With EVALUESTEER, we find even the best RMs we tested exhibit their own value/style biases, and are unable to align with a user &gt;25% of the time. ๐Ÿงต
Letta (@letta_ai) 's Twitter Profile Photo

What if we evaluated agents less like isolated code snippets, and more like humans - where behavior depends on the environment and lived experiences? ๐Ÿงช Introducing ๐—Ÿ๐—ฒ๐˜๐˜๐—ฎ ๐—˜๐˜ƒ๐—ฎ๐—น๐˜€: a fully open source evaluation framework for stateful agents

What if we evaluated agents less like isolated code snippets, and more like humans - where behavior depends on the environment and lived experiences?

๐Ÿงช Introducing ๐—Ÿ๐—ฒ๐˜๐˜๐—ฎ ๐—˜๐˜ƒ๐—ฎ๐—น๐˜€: a fully open source evaluation framework for stateful agents