Armand Joulin (@armandjoulin) 's Twitter Profile
Armand Joulin

@armandjoulin

principal researcher, @googledeepmind. ex director of emea at @metaai. led a few open projects: llama, fasttext, dino, and now gemma.

ID: 22243503

calendar_today28-02-2009 09:38:34

222 Tweet

6,6K Followers

385 Following

Armand Joulin (@armandjoulin) 's Twitter Profile Photo

PaliGemma and Gemini fueled the new vision modality of Gemma 3. It is so great to work at the crossroads of so many amazing teams!

Armand Joulin (@armandjoulin) 's Twitter Profile Photo

Putting open models out there is hard and releases often come with implementation bugs. For Gemma 3, a dedicated release team worked for weeks to avoid these issues. So let s wait before judging llama 4 performances....

Qwen (@alibaba_qwen) 's Twitter Profile Photo

Introducing Qwen3! We release and open-weight Qwen3, our latest large language models, including 2 MoE models and 6 dense models, ranging from 0.6B to 235B. Our flagship model, Qwen3-235B-A22B, achieves competitive results in benchmark evaluations of coding, math, general

Introducing Qwen3! 

We release and open-weight Qwen3, our latest large language models, including 2 MoE models and 6 dense models, ranging from 0.6B to 235B. Our flagship model, Qwen3-235B-A22B, achieves competitive results in benchmark evaluations of coding, math, general
Armand Joulin (@armandjoulin) 's Twitter Profile Photo

we love lmsys and its impact to our field, and thats why we need critical work, like cohere paper to keep improving it. Equally, it is important to clarify that some numbers in Cohere work are inaccurate, e.g., we sent only 1 gemma3 model to lmsys before release ("zizou-10").

Sara Hooker (@sarahookr) 's Twitter Profile Photo

It is critical for scientific integrity that we trust our measure of progress. The lmarena.ai has become the go-to evaluation for AI progress. Our release today demonstrates the difficulty in maintaining fair evaluations on lmarena.ai, despite best intentions.

It is critical for scientific integrity that we trust our measure of progress. 

The <a href="/lmarena_ai/">lmarena.ai</a> has become the go-to evaluation for AI progress.

Our release today demonstrates the difficulty in maintaining fair evaluations on <a href="/lmarena_ai/">lmarena.ai</a>, despite best intentions.