Ruibo Liu (@ruiboliu) 's Twitter Profile
Ruibo Liu

@ruiboliu

RS @GoogleDeepMind. AI Research with Humans in Mind.

ID: 3749425934

linkhttps://ruibo-liu.com/ calendar_today01-10-2015 15:57:27

558 Tweet

2,2K Takipçi

1,1K Takip Edilen

vitrupo (@vitrupo) 's Twitter Profile Photo

Geoffrey Hinton says the more we understand how AI and the brain actually work, the less human thinking looks like logic. We're not reasoning machines, he says. We're analogy machines. We think by resonance, not deduction. “We're much less rational than we thought.”

Ruibo Liu (@ruiboliu) 's Twitter Profile Photo

Very nice work! Congrats, Qwen team! 🥳 And glad to see the comparison includes Gemini 2.5 Pro --- the "best model in the world" in my opinion (yes, even after seeing some new models released these days). More to come... 😊

Ruibo Liu (@ruiboliu) 's Twitter Profile Photo

This is the coolest demo I have seen recently! Now you can do: S_t -> A_t -> S_t+1, R_t ... infinitely! The real world model! 😇

Thang Luong (@lmthang) 's Twitter Profile Photo

Very excited to share that an advanced version of Gemini Deep Think is the first to have achieved gold-medal level in the International Mathematical Olympiad! 🏆, solving five out of six problems perfectly, as verified by the IMO organizers! It’s been a wild run to lead this

Very excited to share that an advanced version of Gemini Deep Think is the first to have achieved gold-medal level in the International Mathematical Olympiad! 🏆, solving five out of six problems perfectly, as verified by the IMO organizers! It’s been a wild run to lead this
Fred Zhang (@fredzhang0) 's Twitter Profile Photo

This is the most scaling-pilled project I've ever been part of, and the team really cooked. TL;DR: With RL and inference scaling, Gemini perfectly solved 5 out of 6 problems, reaching a gold medal in IMO '25, all within the time constraints of 4.5hr.

Ruibo Liu (@ruiboliu) 's Twitter Profile Photo

If the grading is correct then that means we actually don’t need specially trained model or vertical data — the IMO gold level model is already in your hands. Impressive work from the community!

lmarena.ai (formerly lmsys.org) (@lmarena_ai) 's Twitter Profile Photo

🚨 Open Model Leaderboard Update New open models entered the Text Arena, and the rankings by provider have reshuffled for August. - Qwen-3-235b-a22b-instruct from Qwen takes the crown 🏆 - GLM-4.5 from Z.ai and gpt-oss-120b by @openAI debut in the top 10! All the

🚨 Open Model Leaderboard Update

New open models entered the Text Arena, and the rankings by provider have reshuffled for August.

- Qwen-3-235b-a22b-instruct from <a href="/Alibaba_Qwen/">Qwen</a> takes the crown 🏆
- GLM-4.5 from <a href="/Zai_org/">Z.ai</a> and gpt-oss-120b by @openAI debut in the top 10!

All the
Ruibo Liu (@ruiboliu) 's Twitter Profile Photo

There seems to be a strange desperation in the air with AI companies trying to buy web browsers. It’s not just a land grab for distribution; it’s a tacit admission that even with a trillion-parameter model, the user experience is still mediated by a 30-year-old paradigm: the

Jeff Dean (@jeffdean) 's Twitter Profile Photo

VaultGemma is a release of an open model trained from scratch with differential privacy. The blog post below and the full tech report linked from the tech report have some nice analyses to present a scaling law for differentially private language models: Blog:

VaultGemma is a release of an open model trained from scratch with differential privacy.

The blog post below and the full tech report linked from the tech report have some nice analyses to present a scaling law for differentially private language models:

Blog: