Nadav Timor ✈️ ICLR (@keyboardant) 's Twitter Profile
Nadav Timor ✈️ ICLR

@keyboardant

LLM inference, speculative decoding, open source.
Built novel decoding algorithms – default in Hugging Face Transformers (140k+ ⭐). Making LLMs faster + cheaper

ID: 943637648522072065

linkhttps://www.linkedin.com/in/nadav-timor calendar_today21-12-2017 00:22:02

248 Tweet

488 Followers

5,5K Following

Nadav Timor ✈️ ICLR (@keyboardant) 's Twitter Profile Photo

🚀 Calling all OSS contributors to vLLM, SGLang, HF Transformers, TGI, TRT-LLM & more — Join us at the Inference Engines Social @ ICLR 🇸🇬 📍 Singapore EXPO 🗓️ Apr 25, 13:00 Let’s hang out, swap ideas & share what we’re building! 🔗 lu.ma/42pobekx

Nadav Timor ✈️ ICLR (@keyboardant) 's Twitter Profile Photo

🚀 Vercel’s #StateOfAI survey highlights: 83% of teams run ≤3 models, 86% don’t train their own, and 46% say the biggest hurdle is latency or cost (23% latency + 23% cost). Devs are leaning hard on third‑party inference—OpenAI alone powers 83% of workloads

Nadav Timor ✈️ ICLR (@keyboardant) 's Twitter Profile Photo

Hiding the verification latency of speculative decoding = faster inference. Come chat with me about how! 📍 Poster #245 (Hall 3) 🕒 Today, 15:00–17:30

Hiding the verification latency of speculative decoding = faster inference. Come chat with me about how!
📍 Poster #245 (Hall 3)
🕒 Today, 15:00–17:30
Ravid Shwartz Ziv (@ziv_ravid) 's Twitter Profile Photo

Twitter! I can't believe no one told me that people are using vision encoders to retrieve document information these days. Based on a tip from Nadav Timor I read the "ColPali: Efficient Document Retrieval with Vision Language Models" paper, and it is very cool.

Ravid Shwartz Ziv (@ziv_ravid) 's Twitter Profile Photo

Do you want to speed up your inference time by 2X for free? Using speculative decoding (SP) but want to use your own drafter? Want lossless speedup with one line? Go to Nadav Timor talk at ICML tomorrow at 3:30 PM and check out his poster afterward at 4:30 PM. Thank me later!

Do you want to speed up your inference time by 2X for free? Using speculative decoding (SP) but want to use your own drafter? Want lossless speedup with one line?
Go to <a href="/NadavTimor/">Nadav Timor</a>  talk at ICML tomorrow at 3:30 PM and check out his poster afterward at 4:30 PM. Thank me later!
Weizmann Institute (@weizmannscience) 's Twitter Profile Photo

Algorithms developed by researchers in Prof. David Harel’s team and Intel Labs enable AI developers around the world to combine the power of different AI models “thinking” as one. The study led by PhD student Nadav Timor was presented at #ICML2025>> bit.ly/faster-ai-eng

Algorithms developed by researchers in Prof. <a href="/DavidHarel50/">David Harel</a>’s team and Intel Labs enable AI developers around the world to combine the power of different AI models “thinking” as one. The study led by PhD student <a href="/NadavTimor/">Nadav Timor</a> was presented at #ICML2025&gt;&gt; bit.ly/faster-ai-eng