Mehrdad Farajtabar (@mfarajtabar) 's Twitter Profile
Mehrdad Farajtabar

@mfarajtabar

Research Scientist at @Apple, prev @DeepMind, prev @GeorgiaTech

ID: 1346668532210176000

linkhttps://sites.google.com/view/mehrdad calendar_today06-01-2021 04:03:06

147 Tweet

6,6K Takipçi

179 Takip Edilen

Ruoming Pang (@ruomingpang) 's Twitter Profile Photo

In this report we describe the 2025 Apple Foundation Models ("AFM"). We also introduce the new Foundation Models framework, which gives app developers direct access to the on-device AFM model. machinelearning.apple.com/research/apple…

𝚐𝔪𝟾𝚡𝚡𝟾 (@gm8xx8) 's Twitter Profile Photo

Your LLM Knows the Future: Uncovering Its Multi-Token Prediction Potential Autoregressive LMs already know future tokens, this work makes that usable: - Append <mask> tokens → jointly predict k+1 future tokens - Gated LoRA → updates only for MTP tokens, preserving NTP behavior

Your LLM Knows the Future: Uncovering Its Multi-Token Prediction Potential

Autoregressive LMs already know future tokens, this work makes that usable:
- Append &lt;mask&gt; tokens → jointly predict k+1 future tokens
- Gated LoRA → updates only for MTP tokens, preserving NTP behavior
Mehrdad Farajtabar (@mfarajtabar) 's Twitter Profile Photo

It’s great to be excited about AI’s #IMO performance, while also recognizing the true source of its power. I came across this paragraph today during my reading of A Thousand Brains: A New Theory of #Intelligence, 2021, Jeff Hawkins!

It’s great to be excited about AI’s #IMO performance, while also recognizing the true source of its power. I came across this paragraph today during my reading of
A Thousand Brains: A New Theory of #Intelligence, 2021, Jeff Hawkins!
Jackson Atkins (@jacksonatkinsx) 's Twitter Profile Photo

Apple research just revealed a way to make LLMs 5.35x faster. 🤯 That’s not a typo. They've found a method to get a >500% speedup for code & math tasks, with ZERO quality loss. Here's how they're unlocking AI model's "latent potential": 🧵

Apple research just revealed a way to make LLMs 5.35x faster. 🤯

That’s not a typo. They've found a method to get a &gt;500% speedup for code &amp; math tasks, with ZERO quality loss.

Here's how they're unlocking AI model's "latent potential": 🧵
Mehrdad Farajtabar (@mfarajtabar) 's Twitter Profile Photo

I noticed the same thing! Engaging in conversations, replies, or DMs with #DeepMind folks always feels safe and welcoming. Their culture is truly remarkable. Thanks to leaders like Samy Bengio, Devi Krishna, Daphne Luong, JG, and many others who've joined Apple, this incredible

Fartash Faghri (@fartashfg) 's Twitter Profile Photo

📢Submissions are now open for #NeurIPS2025 CCFM workshop. Submission deadline: August 22, 2025, AoE. Website: sites.google.com/view/ccfm-neur… Call for papers: sites.google.com/view/ccfm-neur… Submission Link: openreview.net/group?id=NeurI…

Edward Frenkel (@edfrenkel) 's Twitter Profile Photo

This is an unwise statement that can only make people confused about what LLMs can or cannot do. Let me tell you something: Math is NOT about solving this kind of ad hoc optimization problems. Yeah, by scraping available data and then clustering it, LLMs can sometimes solve some