Jeremy Howard (@jeremyphoward) 's Twitter Profile
Jeremy Howard

@jeremyphoward

🇦🇺 Co-founder: @AnswerDotAI & @FastDotAI ;
Prev: professor @ UQ; Stanford fellow; @kaggle president; @fastmail/@enlitic/etc founder
jeremy.fast.ai

ID: 175282603

linkhttp://answer.ai calendar_today06-08-2010 04:58:18

61,61K Tweet

247,247K Takipçi

5,5K Takip Edilen

Hunter📈🌈📊 (@statisticurban) 's Twitter Profile Photo

- Steve Jobs: son of a Syrian immigrant. - Jeff Bezos: father is a Cuban immigrant. - Sergey Brin: Soviet immigrant, born in Moscow. - Jensen Huang: Taiwanese immigrant. - Elon Musk: South African immigrant. 5 of the magnificent 7.

Casper Hansen (@casper_hansen_) 's Twitter Profile Photo

One of the core limitations of language models: they struggle with true novelty. LLMs remix known information. If something isn’t in the pretraining or finetuning data, they have very little to work with. This is why there are a million papers chasing +5 points on GSM8k.

One of the core limitations of language models: they struggle with true novelty.

LLMs remix known information. If something isn’t in the pretraining or finetuning data, they have very little to work with.

This is why there are a million papers chasing +5 points on GSM8k.
Qwen (@alibaba_qwen) 's Twitter Profile Photo

🚀 Excited to launch Qwen3 models in MLX format today! Now available in 4 quantization levels: 4bit, 6bit, 8bit, and BF16 — Optimized for MLX framework. 👉 Try it now! Huggingface:huggingface.co/collections/Qw… ModelScope: modelscope.cn/collections/Qw…

🚀 Excited to launch Qwen3 models in MLX format today!

Now available in 4 quantization levels: 4bit, 6bit, 8bit, and BF16 — Optimized for MLX framework. 

👉 Try it now!

Huggingface:huggingface.co/collections/Qw…
ModelScope: modelscope.cn/collections/Qw…
Dieter (@kagglingdieter) 's Twitter Profile Photo

“If you want things to be done safely and responsibly, you do it in the open … Don’t do it in a dark room and tell me it’s safe.” Jensen Huang

MiniMax (official) (@minimax__ai) 's Twitter Profile Photo

Day 1/5 of #MiniMaxWeek: We’re open-sourcing MiniMax-M1, our latest LLM — setting new standards in long-context reasoning. - World’s longest context window: 1M-token input, 80k-token output - State-of-the-art agentic use among open-source models - RL at unmatched efficiency:

Day 1/5 of #MiniMaxWeek: We’re open-sourcing MiniMax-M1, our latest LLM — setting new standards in long-context reasoning.

- World’s longest context window: 1M-token input, 80k-token output
- State-of-the-art agentic use among open-source models
- RL at unmatched efficiency:
htmx.org / CEO of div tags (same thing) (@htmx_org) 's Twitter Profile Photo

100% web servers should be gzipping the js on the way out & it should be cached forever afterwards minification destroys debuggability & the (underappreciated) #ViewSource affordance: htmx.org/essays/right-c… /cc Cory Doctorow NONCONSENSUAL BLUE TICK

Charlie Marsh (@charliermarsh) 's Twitter Profile Photo

The Python Steering Council has voted to remove the "experimental" label from the free-threaded ("nogil") builds for Python 3.14. Big step towards making them the default in a future version of CPython!

Jonathan Whitaker (@johnowhitaker) 's Twitter Profile Photo

New video, starting to look at Diffusion Language Models. This one introduces some ideas, then shows how I turn ModernBERT into a LLaDA-style generative model. Lots of avenues to explore from here! Join me in playing with this? Project ideas in thread :) youtube.com/watch?v=Ds_cTc…

Mark Saroufim (@marksaroufim) 's Twitter Profile Photo

Still feels surreal to have been on stage with one of the greatest CEO's of our time Dr Lisa Su where she explicitly called out GPU MODE and the work we did to enable the world's first $100K competitive kernel competition I never in a thousand years would have imagined that a

Still feels surreal to have been on stage with one of the greatest CEO's of our time Dr Lisa Su where she explicitly called out GPU MODE and the work we did to enable the world's first $100K competitive kernel competition

I never in a thousand years would have imagined that a
Alex Zhang (@a1zhang) 's Twitter Profile Photo

btw a shit ton of amazing learning material + open-source code for GPU programming ($150K worth) is linked on the latest GPU MODE news post a year ago when I was an undergrad I was scouring the internet for these kinds of resources, plz take advantage of it!

btw a shit ton of amazing learning material + open-source code for GPU programming ($150K worth) is linked on the latest <a href="/GPU_MODE/">GPU MODE</a> news post

a year ago when I was an undergrad I was scouring the internet for these kinds of resources, plz take advantage of it!
Eric Hartford (@cognitivecompai) 's Twitter Profile Photo

My trick: I copy 2nd LLM's feedback and I tell the first LLM "My drunk friend who is usually wrong, has this feedback:" then I paste the feedback. If the model still agrees with the feedback of drunk usually-wrong friend, then you can trust that it's good feedback.

Oriol Vinyals (@oriolvinyalsml) 's Twitter Profile Photo

Hello Gemini 2.5 Flash-Lite! So fast, it codes *each screen* on the fly (Neural OS concept 👇). The frontier isn't always about large models and beating benchmarks. In this case, a super fast & good model can unlock drastic use cases. Read more: blog.google/products/gemin…

MiniMax (official) (@minimax__ai) 's Twitter Profile Photo

Day 2/5 of #MiniMaxWeek: Introducing Hailuo 02, World-Class Quality, Record-Breaking Cost Efficiency 🎥 - Best-in-class instruction following - Handles extreme physics (yes, it does acrobatics 🤹) - Native 1080p