Daniel Israel (@danielmisrael) 's Twitter Profile
Daniel Israel

@danielmisrael

PhD Student Studying AI/ML @UCLA

ID: 391930949

linkhttps://danielmisrael.github.io/ calendar_today16-10-2011 09:34:20

36 Tweet

821 Takipçi

5,5K Takip Edilen

Anji Liu (@liu_anji) 's Twitter Profile Photo

[1/n] 🚀Diffusion models for discrete data excel at modeling text, but they need hundreds to thousands of diffusion steps to perform well. We show that this is caused by the fact that discrete diffusion models predict each output token *independently* at each denoising step.

[1/n] 🚀Diffusion models for discrete data excel at modeling text, but they need hundreds to thousands of diffusion steps to perform well.

We show that this is caused by the fact that discrete diffusion models predict each output token *independently* at each denoising step.
Christina Chance (@christinachanc) 's Twitter Profile Photo

1/n uclanlp is researching how Black, LGBTQIA+, & women communities perceive and are affected by content moderation, as it relates to English-language social media content using reclaimed language. As part of this, we are recruiting annotators (forms.gle/KP6F9gDCo8Skjs…) …

Zhe Zeng (@zhezeng0908) 's Twitter Profile Photo

📢 I’m recruiting PhD students UVA Computer Science for Fall 2025! 🎯 Neurosymbolic AI, probabilistic ML, trustworthiness, AI for science. See my website for more details: zzeng.me 📬 If you're interested, apply and mention my name in your application: engineering.virginia.edu/department/com…

Benjie Wang (@benjiewang_cs) 's Twitter Profile Photo

You have some model/knowledge (e.g. Bayes Net, Probabilistic/Logic Program, DB) and some query (e.g. MAP, Causal Adjustment) you want to ask. When can you compute this efficiently? Find out @ NeurIPS today in Poster Session 6 East, #3801. Paper: arxiv.org/abs/2412.05481

Tanishq Mathew Abraham, Ph.D. (@iscienceluvr) 's Twitter Profile Photo

Enabling Autoregressive Models to Fill In Masked Tokens Hybrid autoregressive and masked language model for infilling by training a linear decoder that takes their concatenated hidden states as input. Provides faster inference with KV caching. MARIA significantly outperforms

Enabling Autoregressive Models to Fill In Masked Tokens

Hybrid autoregressive and masked language model for infilling by training a linear decoder that takes their concatenated hidden states as input. Provides faster inference with KV caching. MARIA significantly outperforms
Siyan Zhao (@siyan_zhao) 's Twitter Profile Photo

Excited to release PrefEval (ICLR '25 Oral), a benchmark for evaluating LLMs’ ability to infer, memorize, and adhere to user preferences in long-context conversations! ⚠️We find that cutting-edge LLMs struggle to follow user preferences—even in short contexts. This isn't just

Excited to release PrefEval (ICLR '25 Oral), a benchmark for evaluating LLMs’ ability to infer, memorize, and adhere to user preferences in long-context conversations!

⚠️We find that cutting-edge LLMs struggle to follow user preferences—even in short contexts. This isn't just
Aditya Grover (@adityagrover_) 's Twitter Profile Photo

A few months ago, we started Inception Labs, a new generative AI startup with a rockstar founding team. At Inception, we are challenging the status quo for language generation. Our first results bring blazing fast speeds at 1000+ tokens/sec while matching the quality of leading

Hritik Bansal (@hbxnov) 's Twitter Profile Photo

Video generative models hold the promise of being general-purpose simulators of the physical world 🤖 How far are we from this goal❓ 📢Excited to announce VideoPhy-2, the next edition in the series to test the physical likeness of the generated videos for real-world actions. 🧵

Video generative models hold the promise of being general-purpose simulators of the physical world 🤖 How far are we from this goal❓

📢Excited to announce VideoPhy-2, the next edition in the series to test the physical likeness of the generated videos for real-world actions. 🧵
Zilei Shao (@zileishao) 's Twitter Profile Photo

What happens if we tokenize cat as [ca, t] rather than [cat]? LLMs are trained on just one tokenization per word, but they still understand alternative tokenizations. We show that this can be exploited to bypass safety filters without changing the text itself. #AI #LLMs #Token

What happens if we tokenize cat as [ca, t] rather than [cat]? 

LLMs are trained on just one tokenization per word, but they still understand alternative tokenizations. We show that this can be exploited to bypass safety filters without changing the text itself.

#AI #LLMs #Token
Hritik Bansal (@hbxnov) 's Twitter Profile Photo

📢Scaling test-time compute via generative verification (GenRM) is an emerging paradigm and shown to be more efficient than self-consistency (SC) for reasoning. But, such claims are misleading☠️ Our compute-matched analysis shows that SC outperforms GenRM across most budgets! 🧵

📢Scaling test-time compute via generative verification (GenRM) is an emerging paradigm and shown to be more efficient than self-consistency (SC) for reasoning. But, such claims are misleading☠️

Our compute-matched analysis shows that SC outperforms GenRM across most budgets! 🧵
Lucas Bandarkar (@lucasbandarkar) 's Twitter Profile Photo

The unreasonable effectiveness of model merging for cross-lingual transfer ! Our preprint evaluates a number of *modular* approaches to fine-tuning LLMs that "assign" model params to either task or language. Surprisingly, merging experts beats all ! 🧵1/4 arxiv.org/abs/2505.18356