Mick (@mickvanhulst) 's Twitter Profile
Mick

@mickvanhulst

Staff Applied scientist @TomTom working on automated mapmaking (HD Maps). Opinions my own.

ID: 904051307350196224

linkhttps://github.com/mickvanhulst calendar_today02-09-2017 18:40:02

341 Tweet

100 Followers

1,1K Following

Bingyi Kang (@bingyikang) 's Twitter Profile Photo

After a year of team work, we're thrilled to introduce Depth Anything 3 (DA3)! 🚀 Aiming for human-like spatial perception, DA3 extends monocular depth estimation to any-view scenarios, including single images, multi-view images, and video. In pursuit of minimal modeling, DA3

Yunhao Luo (@yluo_y) 's Twitter Profile Photo

🚀Excited to share our NeurIPS Conference paper Compositional Diffuser (CompDiffuser)! CompDiffuser scales planning horizons at test time, able to construct long-horizon plans while only trained on short-horizon data. Project page: comp-diffuser.github.io Thread 👇(1/n)

Chieh-Hsin (Jesse) Lai (@jcjesselai) 's Twitter Profile Photo

🎉 Our diffusion-model monograph 《The Principles of Diffusion Models》 now has an official website! 🔗 Link in thread 💡 Two highlights: 1️⃣ Blog Post: a big-picture walkthrough from diffusion models → fast generation & flow map models; 2️⃣ Teaching Guide: lightweight course

sway (@swaystar123) 's Twitter Profile Photo

To everyone who was questioning the muon results: You are right! The LR was badly tuned, and weight decay was also hampering it. Using lr 1e-3 and 0 wd, got a new best, 3.39 FID @ 400k

Mick (@mickvanhulst) 's Twitter Profile Photo

Whenever I read about LLMs as a jury, I wonder how different these LLMs really are. Ideally, one would assume their errors are uncorrelated, so that if their mistakes are independent, the error rate becomes epsilon squared. But their base models have all been trained on the

Hamel Husain (@hamelhusain) 's Twitter Profile Photo

This deserved its own flashcard b/c I've seen bus stop ads from eval vendors encouraging the opposite in San Francisco 🤣 The only thing generic metrics do is waste your time. Links in reply.

This deserved its own flashcard b/c I've seen bus stop ads from eval vendors encouraging the opposite in San Francisco 🤣

The only thing generic metrics do is waste your time.  Links in reply.
Niels Rogge (@nielsrogge) 's Twitter Profile Photo

One of the best visual explanations I've ever seen for why scaling Transformers works, but is suboptimal, as it's just brute-forcing things, by Llion Jones (co-author of the Transformer) on Machine Learning Street Talk "In the (rejected) paper "Intelligent Matrix Exponentiation", they show

One of the best visual explanations I've ever seen for why scaling Transformers works, but is suboptimal, as it's just brute-forcing things, by <a href="/YesThisIsLion/">Llion Jones</a> (co-author of the Transformer) on <a href="/MLStreetTalk/">Machine Learning Street Talk</a> 

"In the (rejected) paper "Intelligent Matrix Exponentiation", they show
Mick (@mickvanhulst) 's Twitter Profile Photo

Quite curious about Grokipedia! May be a great opportunity to make it a verifiable large-scale knowledge graph for better context management via graph traversal. Reduces LLM need to memorize the world, enabling a "cognitive core"-like approach (see Karpathy's tweets). Do wonder

Google (@google) 's Twitter Profile Photo

We’re launching full-length, on demand practice exams for standardized tests in Google Gemini, starting with the SAT, available now at no cost. Practice SATs are grounded in rigorously vetted content in partnership with The Princeton Review, and Gemini will provide immediate feedback

Wall Street Mav (@wallstreetmav) 's Twitter Profile Photo

If science class explained everything with Midwest emo music, we would have a lot more people paying attention. Nuclear power explained. It’s just a fancy way to heat water and make electricity. 🔊

Boring_Business (@boringbiz_) 's Twitter Profile Photo

This was an eye opener from Jensen Huang When asked whether he would rather relive his 20s or be 20 years old today, this is what he had to say: "I thought our 20s were happier than these 20s. I think everyone deserves some time to be oblivious, and not wear all of the world's

Robert Cincotta (@drrobcincotta) 's Twitter Profile Photo

I have been testing the new Obsidian CLI with Claude Code on my research vault (4,663 files, 16 GB)... I know too many notes!! Early results are significant. Its going to change the way in which Claude Code can interact with Obsidian The way I see it, there are three ways

I have been testing the new Obsidian CLI with Claude Code on my research vault (4,663 files, 16 GB)... I know too many notes!!

Early results are significant.  Its going to change the way in which Claude Code can interact with Obsidian

The way I see it, there are three ways
Dimitri (@dimitrikennedy) 's Twitter Profile Photo

The next big level up for me wont be a better model, just boring infra. Better container orchestration. Personal compute networking. Secret management. Environment replication. Designed for a world where one person is running 100+ parallel workstreams.

Om Patel (@om_patel5) 's Twitter Profile Photo

I taught Claude to talk like a caveman to use 75% less tokens. normal claude: ~180 tokens for a web search task caveman claude: ~45 tokens for the same task "I executed the web search tool" = 8 tokens caveman version: "Tool work" = 2 tokens every single grunt swap saves 6-10

I taught Claude to talk like a caveman to use 75% less tokens.

normal claude: ~180 tokens for a web search task

caveman claude: ~45 tokens for the same task

"I executed the web search tool" = 8 tokens
caveman version: "Tool work" = 2 tokens

every single grunt swap saves 6-10
James Zou (@james_y_zou) 's Twitter Profile Photo

Big Update🤩: #paperclip now includes full papers from all of arXiv, PubMed Central and 150 million abstracts!🖇️ You can give your LLM all that knowledge in one line—all optimally indexed for AI agents. Much more thorough and ~100x faster than web search, and free.

Big Update🤩: #paperclip now includes full papers from all of arXiv, PubMed Central and 150 million abstracts!🖇️

You can give your LLM all that knowledge in one line—all optimally indexed for AI agents. Much more thorough and ~100x faster than web search, and free.