Nikolay Savinov 🇺🇦 (@savinovnikolay) 's Twitter Profile
Nikolay Savinov 🇺🇦

@savinovnikolay

Research Scientist at @GoogleDeepMind
Work on LLM pre-training in Gemini ♊
Lead 10M context length in Gemini 1.5 📈

ID: 973916582530420736

linkhttps://www.nsavinov.com calendar_today14-03-2018 13:39:42

876 Tweet

2,2K Takipçi

0 Takip Edilen

Google DeepMind (@googledeepmind) 's Twitter Profile Photo

We’re releasing an updated Gemini 2.5 Pro (I/O edition) to make it even better at coding. 🚀 You can build richer web apps, games, simulations and more - all with one prompt. In Google Gemini App, here's how it transformed images of nature into code to represent unique patterns 🌱

Logan Kilpatrick (@officiallogank) 's Twitter Profile Photo

Gemini 2.5 Pro just got an upgrade & is now even better at coding, with significant gains in front-end web dev, editing, and transformation. We also fixed a bunch of function calling issues that folks have been reporting, it should now be much more reliable. More details in 🧵

Gemini 2.5 Pro just got an upgrade & is now even better at coding, with significant gains in front-end web dev, editing, and transformation.

We also fixed a bunch of function calling issues that folks have been reporting, it should now be much more reliable. More details in 🧵
lmarena.ai (formerly lmsys.org) (@lmarena_ai) 's Twitter Profile Photo

All these category strengths by the latest Gemini-2.5-Pro. 💻 Coding, Math, Creative Writing, Longer Query, ... the community loved it across all categories.

All these category strengths by the latest Gemini-2.5-Pro.

💻 Coding, Math, Creative Writing, Longer Query, ... the community loved it across all categories.
Jeffrey Emanuel (@doodlestein) 's Twitter Profile Photo

The new update to Gemini 2.5 Pro is awesome. It’s totally incredible to me that I can now dump 2 megabytes of code (36.5k lines of mostly Python and some HTML/JS) and it can do a really great job understanding everything and helping me. The conversation STARTS with 470k tokens!

Ethan Mollick (@emollick) 's Twitter Profile Photo

Pretty awesome result from the new version of Gemini 2.5 I changed one line of War and Peace, inserting a sentence into Book 14, Chapter 10 (halfway through), where Princess Mary "spoke to Crab Man the superhero" Gemini 2.5 consistently found this reference among 860,000 tokens

Pretty awesome result from the new version of Gemini 2.5

I changed one line of War and Peace, inserting a sentence into Book 14, Chapter 10 (halfway through), where Princess Mary "spoke to Crab Man the superhero"

Gemini 2.5 consistently found this reference among 860,000 tokens
Deedy (@deedydas) 's Twitter Profile Photo

1. Take a screen recording explaining your app 2. Upload it to YouTube 3. "Build me this” Gemini 2.5’s ability to comprehend video feels straight out of a science fiction novel.

Nikolay Savinov 🇺🇦 (@savinovnikolay) 's Twitter Profile Photo

No need to explicitly specify caching in Gemini API anymore, now it's done automatically - important for long-context applications, e.g. coding!

Yam Peleg (@yampeleg) 's Twitter Profile Photo

Gemini’s attention to detail is wild. Fed a ~400k token codebase and forgot about it Found the tab still open a day later Just for fun, I dumped the entire project again after refactoring it for a full day and asked it to guess my goal

Jeff Dean (@jeffdean) 's Twitter Profile Photo

Caching of tokens when you use the same input context repeatedly now happens implicitly, making things a lot easier to deal with.

Andriy Burkov (@burkov) 's Twitter Profile Photo

Crazy times. This morning, while getting breakfast, I instructed Gemini to write a Python script that connects to my Spotify, pulls my liked songs, and uses Gemini to recommend more songs. Each time I run the script, my Spotify playlist gets updated with 20 new recommended songs

lmarena.ai (formerly lmsys.org) (@lmarena_ai) 's Twitter Profile Photo

🚨Breaking from Arena: Google DeepMind's new Gemini-2.5-Flash climbs to #2 overall in chat, a major jump from its April release (#5 → #2)! Highlights: - Top-2 across major categories (Hard, Coding, Math) - #3 in WebDev Arena, #2 in Vision Arena - New model at the

🚨Breaking from Arena: <a href="/GoogleDeepMind/">Google DeepMind</a>'s new Gemini-2.5-Flash climbs to #2 overall in chat, a major jump from its April release (#5 → #2)!

Highlights:
- Top-2 across major categories (Hard, Coding, Math)
- #3 in WebDev Arena, #2 in Vision Arena
- New model at the
Logan Kilpatrick (@officiallogank) 's Twitter Profile Photo

The new Gemini 2.5 Pro is SOTA at long context, especially capable on higher number of items being retrieved (needles) as shown below!

The new Gemini 2.5 Pro is SOTA at long context, especially capable on higher number of items being retrieved (needles) as shown below!
Paul Gauthier (@paulgauthier) 's Twitter Profile Photo

Gemini 2.5 Pro 06-05 has set a new SOTA on the aider polyglot coding benchmark, scoring 83% with 32k thinking tokens. The default thinking mode, where Gemini self-determines the thinking budget, scored 79%. Full leaderboard: aider.chat/docs/leaderboa…

Gemini 2.5 Pro 06-05 has set a new SOTA on the aider polyglot coding benchmark, scoring 83% with 32k thinking tokens.

The default thinking mode, where Gemini self-determines the thinking budget, scored 79%.

Full leaderboard:
aider.chat/docs/leaderboa…
Deedy (@deedydas) 's Twitter Profile Photo

China's Gaokao is the biggest exam in the world: 13M test takers and 9hrs. ~0.02% make it to the top uni, Tsinghua. As of this week, AI models can make it too. 625/750 is top 1%ile. Highest human score is ~720-740. Gemini 2.5 Pro gets 655, barely making the cut for Tsinghua!

China's Gaokao is the biggest exam in the world: 13M test takers and 9hrs. ~0.02% make it to the top uni, Tsinghua. As of this week, AI models can make it too.

625/750 is top 1%ile. Highest human score is ~720-740.

Gemini 2.5 Pro gets 655, barely making the cut for Tsinghua!