Pranav Nair (@pranavn1008) 's Twitter Profile
Pranav Nair

@pranavn1008

Predoctoral Researcher @ Google DeepMind

ID: 1509787001645862912

linkhttps://pranavajitnair.github.io/ calendar_today01-04-2022 06:57:54

24 Tweet

457 Followers

248 Following

yobibyte (@y0b1byte) 's Twitter Profile Photo

Jeff shows a great example on how a senior author presents their contribution! 'Minor co-author' as opposed to popular 'Equal Contribution Senior Co-Adviser.

Prateek Jain (@jainprateek_) 's Twitter Profile Photo

Super excited about the new MatQuant work! Allows training a quantized model where 2bit weights are nested within 4bits and so on. This enables "reading" off accurate models that can have 2bit quantization in the first layer, 4bit in the second layer etc. Along with the

Sachin Yadav (@sachinyv) 's Twitter Profile Photo

โœจNew Paper: Presenting Interleaved Gibbs Diffusion (IGD), a novel generative framework for mixed continuous-discrete data, focusing on constrained generation. From 3-SAT and molecule design to layout generation, IGD advances diffusion models by capturing complex inter-variable

Google DeepMind (@googledeepmind) 's Twitter Profile Photo

Think you know Gemini? ๐Ÿค” Think again. Meet Gemini 2.5: our most intelligent model ๐Ÿ’ก The first release is Pro Experimental, which is state-of-the-art across many benchmarks - meaning it can handle complex problems and give more accurate responses. Try it now โ†’

lmarena.ai (formerly lmsys.org) (@lmarena_ai) 's Twitter Profile Photo

Gemini 2.5 Pro #1 across ALL categories, tied #1 with Grok-3/GPT-4.5 for Hard Prompts and Coding, and edged out across all others to take the lead ๐Ÿ‡๐Ÿ†

Gemini 2.5 Pro #1 across ALL categories, tied #1 with Grok-3/GPT-4.5 for Hard Prompts and Coding, and edged out across all others to take the lead ๐Ÿ‡๐Ÿ†
Zain (@zainhasan6) 's Twitter Profile Photo

First Video from the Learning Together Series on Matryoshka machine learning is live now! Aditya covered everything on matryoshka starting with embeddings, transformers and quantization.

First Video from the Learning Together Series on Matryoshka machine learning is live now!

Aditya covered everything on matryoshka starting with embeddings, transformers and quantization.
lmarena.ai (formerly lmsys.org) (@lmarena_ai) 's Twitter Profile Photo

๐ŸšจBreaking: Google DeepMindโ€™s latest Gemini-2.5-Pro is now ranked #1 across all LMArena leaderboards ๐Ÿ† Highlights: - #1 in all text arenas (Coding, Style Control, Creative Writing, etc) - #1 on the Vision leaderboard with a ~70 pts lead! - #1 on WebDev Arena, surpassing Claude

๐ŸšจBreaking: <a href="/GoogleDeepMind/">Google DeepMind</a>โ€™s latest Gemini-2.5-Pro is now ranked #1 across all LMArena leaderboards ๐Ÿ†

Highlights:
- #1 in all text arenas (Coding, Style Control, Creative Writing, etc)
- #1 on the Vision leaderboard with a ~70 pts lead!
- #1 on WebDev Arena, surpassing Claude
Google DeepMind (@googledeepmind) 's Twitter Profile Photo

Weโ€™ve developed Gemini Diffusion: our state-of-the-art text diffusion model. Instead of predicting text directly, it learns to generate outputs by refining noise, step-by-step. This helps it excel at coding and math, where it can iterate over solutions quickly. #GoogleIO

Aditya Kusupati (@adityakusupati) 's Twitter Profile Photo

Pocket powerhouse admist I/O awesomeness! Gemma 3n E4B & E2B are insane models, optimized for on-device while rivaling frontier models. It's a ๐Ÿช†Matryoshka Transformer (MatFormer)๐Ÿช†: Natively elastic b/w 4B & 2B pareto-optimally! โญ๏ธ: free models with ZERO training cost! ๐Ÿงต๐Ÿ‘‡

Pranav Nair (@pranavn1008) 's Twitter Profile Photo

Interesting work on reducing reward hacking. Trains a reward model that is aware of the causal attributes pertaining to evaluation.

Aditya Kusupati (@adityakusupati) 's Twitter Profile Photo

๐Ÿ“ขNow open, Gemma 3n weights & it is natively flexible, first of its kind, thanks to MatFormer๐Ÿช† Any model between E4B & E2B with ZERO training near Pareto -- we found a bunch! Find a better E3B than what we released, I will send you a ๐Ÿช†๐Ÿ˜‰ Find the colab for extraction ๐Ÿงต๐Ÿ‘‡๐Ÿช†

๐Ÿ“ขNow open, Gemma 3n weights &amp; it is natively flexible, first of its kind, thanks to MatFormer๐Ÿช†

Any model between E4B &amp; E2B with ZERO training near Pareto -- we found a bunch!

Find a better E3B than what we released, I will send you a ๐Ÿช†๐Ÿ˜‰

Find the colab for extraction ๐Ÿงต๐Ÿ‘‡๐Ÿช†
Sahil Goyal (@sahilgo6801) 's Twitter Profile Photo

Hi, we'll be presenting MaGNeTS (arxiv.org/pdf/2502.00382) on 15th July at #ICML2025 ๐Ÿ“East Exhibition Hall A-B #3209 ๐Ÿ•ฆ 11 AM - 1:30PM Excited to discuss about nested transformers and decode time scaling for visual generation!

Hi, we'll be presenting  MaGNeTS (arxiv.org/pdf/2502.00382)  on 15th July at #ICML2025 
๐Ÿ“East Exhibition Hall A-B #3209
๐Ÿ•ฆ 11 AM - 1:30PM

Excited to discuss about nested transformers and decode time scaling for visual generation!
Prateek Jain (@jainprateek_) 's Twitter Profile Photo

Puranjay will present our poster on nested bitwise models or MatQuant, so if you are ICML and interested in the topic, do bother him :) Puranjay is going on the grad-school market this cycle. So if you are looking for a brilliant, hardworking student with good ML+LLM exposure,

Aditya Kusupati (@adityakusupati) 's Twitter Profile Photo

๐Ÿช† Matryoshka is extremely general & applicable to every component in our modern ML/DL stack. It can't get more fundamental that ๐Ÿช† in bit space to enable elastic quantization! Drop by the poster and say hi to Puranjay (on behalf of Pranav Nair Jeff Dean Prateek Jain & me).