Agentica Project (@agentica_) 's Twitter Profile
Agentica Project

@agentica_

Building generalist agents that scale @BerkeleySky

ID: 1884497281870929920

linkhttp://www.agentica-project.com calendar_today29-01-2025 07:02:25

55 Tweet

2,2K Followers

8 Following

Sijun Tan (@sijun_tan) 's Twitter Profile Photo

Check out Michael Luo's latest work, Autellix—an ultra-fast system for serving agentic workloads, achieving 4-15x speedups over vLLM/SGLang! At Agentica, we are committed to building efficient infra for serving/training of LLM agents, and Autellix is the first step towards it!

Agentica Project (@agentica_) 's Twitter Profile Photo

🚀Just two weeks after we open-sourced our DeepScaleR model, and the community has already reproduced our results - surpassing O1-preview by following our training recipe! 🎉 Our goal has always been to democratize RL training for LLMs, and by sharing everything—data, code, and

🚀Just two weeks after we open-sourced our DeepScaleR model, and the community has already reproduced our results - surpassing O1-preview by following our training recipe! 🎉

Our goal has always been to democratize RL training for LLMs, and by sharing everything—data, code, and
Together AI (@togethercompute) 's Twitter Profile Photo

Announcing DeepCoder-14B – an o1 & o3-mini level coding reasoning model fully open-sourced! We’re releasing everything: dataset, code, and training recipe.🔥 Built in collaboration with the Agentica Project team. See how we created it. 🧵

Announcing DeepCoder-14B – an o1 & o3-mini level coding reasoning model fully open-sourced!

We’re releasing everything: dataset, code, and training recipe.🔥

Built in collaboration with the <a href="/Agentica_/">Agentica Project</a> team.

See how we created it. 🧵
𝚐𝔪𝟾𝚡𝚡𝟾 (@gm8xx8) 's Twitter Profile Photo

DeepCoder-14B is a code reasoning LLM fine-tuned from DeepSeek-R1-Distill-Qwen-14B using distributed RL with GRPO+ and iterative context lengthening. Trained on ~24K coding problems (TACO-Verified, PrimeIntellect SYNTHETIC-1, LCB v5), it improves Pass@1 on LiveCodeBench v5 to

DeepCoder-14B is a code reasoning LLM fine-tuned from DeepSeek-R1-Distill-Qwen-14B using distributed RL with GRPO+ and iterative context lengthening. Trained on ~24K coding problems (TACO-Verified, PrimeIntellect SYNTHETIC-1, LCB v5), it improves Pass@1 on LiveCodeBench v5 to
Michael Luo (@michaelzluo) 's Twitter Profile Photo

🚀 We introduce DeepCoder-14B-Preview, a fully open-sourced coding model that is on par with o3-mini and o1! 📷 We scaled our model with RL magic up to 32K context. It's performance scales to 64K context 🔥

🚀 We introduce DeepCoder-14B-Preview, a fully open-sourced coding model that is on par with o3-mini and o1!

📷 We scaled our model with RL magic up to 32K context.  It's performance scales to 64K context 🔥
Sijun Tan (@sijun_tan) 's Twitter Profile Photo

Hey Sam Altman, we know you're planning to open-source your reasoning model—but we couldn’t wait. Introducing DeepCoder-14B-Preview: a fully open-source reasoning model that matches o1 and o3-mini on both coding and math. And yes, we’re releasing everything: model, data, code, and

Roy Huang (@_royh021) 's Twitter Profile Photo

😱14B Fully Open Source o3-mini level model?? 😱 DeepCoder is a small but mighty reasoning model that is o3-mini and o1 level on coding and math that runs at a tiny fraction of the cost. Another W by open source. Go check it out!!

Yuchen Jin (@yuchenj_uw) 's Twitter Profile Photo

UC Berkeley open-sourced a 14B model that rivals OpenAI o3-mini and o1 on coding! They applied RL to Deepseek-R1-Distilled-Qwen-14B on 24K coding problems. It only costs 32 H100 for 2.5 weeks (~$26,880)! It's truly open-source. They released everything: the model, training

UC Berkeley open-sourced a 14B model that rivals OpenAI o3-mini and o1 on coding!

They applied RL to Deepseek-R1-Distilled-Qwen-14B on 24K coding problems.

It only costs 32 H100 for 2.5 weeks (~$26,880)!

It's truly open-source. They released everything: the model, training