Yannis Flet-Berliac (@yfletberliac) 's Twitter Profile
Yannis Flet-Berliac

@yfletberliac

Post-training & RL Research @cohere | Postdoc @Stanford | PhD @Inria (Sequel/Scool) | Associate Program Chair @icmlconf’23

ID: 1465200073

linkhttps://cs.stanford.edu/~yfletberliac calendar_today28-05-2013 17:00:10

334 Tweet

760 Followers

950 Following

Cohere Labs (@cohere_labs) 's Twitter Profile Photo

Everyone is on their way to Vancouver for NeurIPS Conference 2024! ✈️ We are looking forward to connecting with everyone! 🤗 Here is where you can find us:

Everyone is on their way to Vancouver for <a href="/NeurIPSConf/">NeurIPS Conference</a> 2024! ✈️

We are looking forward to connecting with everyone! 🤗
Here is where you can find us:
aakanksha (@____aakanksha) 's Twitter Profile Photo

Started the morning off with the BEST pain au chocolat and views in Vancouver for #NeurIPS2024 🥐🏔️ Presenting this work at the Safe Generative AI workshop on Sunday, December 15! I’ll also be at the Cohere booth later today to chat all things model merging, safety and more! 💫

Allen Nie (🇺🇦☮️) (@allen_a_nie) 's Twitter Profile Photo

Time to prove that I'm not all about that LLM bandwagon -- presenting work w/ Yash Chandak Yannis Flet-Berliac, Christina, Ani, and Emma Brunskill on Offline RL policy evaluation. Friday Dec 13: 11am-2 pm Poster Location: West Ballroom A-D Stanford AI Lab Stanford HAI

Time to prove that I'm not all about that LLM bandwagon -- presenting work w/  <a href="/chandakyash13/">Yash Chandak</a> <a href="/yfletberliac/">Yannis Flet-Berliac</a>, Christina, Ani, and <a href="/EmmaBrunskill/">Emma Brunskill</a> on Offline RL policy evaluation. 
Friday Dec 13: 11am-2 pm
Poster Location: West Ballroom A-D 
<a href="/StanfordAILab/">Stanford AI Lab</a> <a href="/StanfordHAI/">Stanford HAI</a>
Yannis Flet-Berliac (@yfletberliac) 's Twitter Profile Photo

📬 This is what we've been cooking the last few months - check out our new 7B model! Smaller and faster, it beats Gemma-2 9B, Ministral 8B and Llama-3.1 8B on the OpenLLM leaderboard. It has best-in-class RAG, tool use, agents + enhanced efficiency in reasoning, maths and code.

📬 This is what we've been cooking the last few months - check out our new 7B model!

Smaller and faster, it beats Gemma-2 9B, Ministral 8B and Llama-3.1 8B on the OpenLLM leaderboard.

It has best-in-class RAG, tool use, agents + enhanced efficiency in reasoning, maths and code.
Daniel San (@dani_avila7) 's Twitter Profile Photo

Trying out Command R7B in VSCode, and the model performs brilliantly! 👏 The latest model from cohere Command family shows excellent performance working with code inside VSCode, using CodeGPT to integrate the model. Congrats to the Cohere team! 🥳 If you want to use Cohere's

Command A(idan) (@aidangomez) 's Twitter Profile Photo

I’m so excited to share something we’ve been working on for awhile: North is cohere’s AI workspace for enterprises. Today we’re releasing the platform for early access!

I’m so excited to share something we’ve been working on for awhile: North is <a href="/cohere/">cohere</a>’s AI workspace for enterprises. Today we’re releasing the platform for early access!
Cohere Labs (@cohere_labs) 's Twitter Profile Photo

Introducing ✨ Aya Vision ✨ - an open-weights model to connect our world through language and vision Aya Vision adds breakthrough multimodal capabilities to our state-of-the-art multilingual 8B and 32B models. 🌿

cohere (@cohere) 's Twitter Profile Photo

We’re excited to introduce our newest state-of-the-art model: Command A! Command A provides enterprises maximum performance across agentic tasks with minimal compute requirements.

Yannis Flet-Berliac (@yfletberliac) 's Twitter Profile Photo

Excited to share Command A! It outperforms GPT-4o and DeepSeek-V3 with unmatched efficiency. It excels in business, STEM, and coding tasks while being 1.75x faster than GPT-4o and 2.4x faster than DeepSeek-V3. Glad to have contributed to its post-training with a fabulous team!

Excited to share Command A!

It outperforms GPT-4o and DeepSeek-V3 with unmatched efficiency. It excels in business, STEM, and coding tasks while being 1.75x faster than GPT-4o and 2.4x faster than DeepSeek-V3.

Glad to have contributed to its post-training with a fabulous team!
Cohere Labs (@cohere_labs) 's Twitter Profile Photo

Following the open-weight release of Command A and Command R7B models, we're excited to have collaborated with @Cohere colleagues on a tech report highlighting our novel approach to model training, including self-refinement algorithms and model merging techniques at scale.

Following the open-weight release of Command A and Command R7B models, we're excited to have collaborated with @Cohere colleagues on a tech report highlighting our novel approach to model training, including self-refinement algorithms and model merging techniques at scale.
Yannis Flet-Berliac (@yfletberliac) 's Twitter Profile Photo

Excited to finally share that CoPG — the RL method I co-authored with Nathan Grinsztajn and amazing colleagues — was used throughout the post-training (offline & online learning) of cohere’s new Command models! 🖊️ Tech report: cohere.com/research/paper… 🤖 CoPG: arxiv.org/abs/2406.19185

Excited to finally share that CoPG — the RL method I co-authored with <a href="/NGrinsztajn/">Nathan Grinsztajn</a> and amazing colleagues — was used throughout the post-training (offline &amp; online learning) of <a href="/cohere/">cohere</a>’s new Command models!

🖊️ Tech report: cohere.com/research/paper…
🤖 CoPG: arxiv.org/abs/2406.19185
Yannis Flet-Berliac (@yfletberliac) 's Twitter Profile Photo

Would you like to use Q-learning for LLM fine-tuning? Check out our new preprint where we interpret Q-functions as logits of the LLM: arxiv.org/abs/2505.11081 ✨ Work done with my based colleagues cohere