Lianmin Zheng (@lm_zheng) 's Twitter Profile
Lianmin Zheng

@lm_zheng

Member of technical staff @xAI | Prev: Ph.D. @UCBerkeley, Co-founder @lmsysorg

ID: 952696452173594625

linkhttp://lmzheng.net/ calendar_today15-01-2018 00:18:29

343 Tweet

11,11K Takipçi

563 Takip Edilen

verl project (@verl_project) 's Twitter Profile Photo

We will present latest updates of verl at #ICLR2025: - recent RL recipes (DAPO, etc) - RL with tool calling & multi-turn - full sglang integration (with LMSYS Org ) - large scale optimizations, and many more Come join us!

We will present latest updates of verl at #ICLR2025:
- recent RL recipes (DAPO, etc)
- RL with tool calling & multi-turn
- full sglang integration (with <a href="/lmsysorg/">LMSYS Org</a> )
- large scale optimizations, and many more

Come join us!
LMSYS Org (@lmsysorg) 's Twitter Profile Photo

Excited to co-host the "Frontiers of Generative AI" afterparty with Abaka AI at #ICLR2025! 🚀 Come connect with fellow researchers and enthusiasts from LMSYS Org and beyond. 🗓️ When: April 26th (Sat), 18:30 - 21:30 🗺️ Where: 10-min from Singapore EXPO (Register for exact details)

Excited to co-host the "Frontiers of Generative AI" afterparty with Abaka AI at #ICLR2025! 🚀
Come connect with fellow researchers and enthusiasts from LMSYS Org and beyond.
🗓️ When: April 26th (Sat), 18:30 - 21:30
🗺️ Where: 10-min from Singapore EXPO (Register for exact details)
LMSYS Org (@lmsysorg) 's Twitter Profile Photo

The SkyPilot team's benchmark for SGLang and vLLM shows that SGLang consistently outperforms with a 30% performance advantage. We will continue to optimize. Cheers!

Ebby Amir (@ebbyamir) 's Twitter Profile Photo

Introducing Grok Vision, multilingual audio, and realtime search in Voice Mode. Available now. Grok habla español Grok parle français Grok Türkçe konuşuyor グロクは日本語を話す ग्रोक हिंदी बोलता है

lmarena.ai (formerly lmsys.org) (@lmarena_ai) 's Twitter Profile Photo

Around this time 2 years ago, the community helped us launch the very first Arena leaderboard! Today we’re publishing a blog to celebrate everything we’ve built together on LMArena! 🥳👏 Highlights: ☑️ 3M+ community votes 🤖 400+ models ranked across text, vision,

Around this time 2 years ago, the community helped us launch the very first Arena leaderboard!

Today we’re publishing a blog to celebrate everything we’ve built together on LMArena! 🥳👏

Highlights:
☑️ 3M+ community votes
🤖 400+ models ranked across text, vision,
Junyang Lin (@justinlin610) 's Twitter Profile Photo

LMSYS Org thanks to the support of SGL! They have been dedicated to helping us deploy the large MoE models and improve the inference efficiency. 🌊

<a href="/lmsysorg/">LMSYS Org</a> thanks to the support of SGL! They have been dedicated to helping us deploy the large MoE models and improve the inference efficiency. 🌊
Biao He (@hebiao064) 's Twitter Profile Photo

📷  Thrilled to share my latest blog post  (co-authored with Qingquan (QQ) Song ) on our open source work integrating the Flash Attention Backend end-to-end in SGLang—now enabled by default in the latest version! 😎 Dive into the details here: hebiao064.github.io/fa3-attn-backe… #LLM #SGLang

Hao AI Lab (@haoailab) 's Twitter Profile Photo

🚨 New Challenger: GROK joins the Game Arena Benchmark! We evaluated Grok3-mini-beta: thinkining on four games: 🧩 2048 | 🧱 Sokoban | 🍬 Candy Crush | 🎮 Phoenix Wright With fast progress, it’s already comparable to top models like OpenAI’s O1, previous O3-mini, and

LMSYS Org (@lmsysorg) 's Twitter Profile Photo

🚀 Breaking: SGLang provides the first open-source implementation to serve DeepSeek V3/R1 models with large-scale expert parallelism and prefill-decode disaggregation on 96 GPUs. It nearly matches the throughput reported by the official DeepSeek blog, achieving 52.3K input

🚀 Breaking: SGLang provides the first open-source implementation to serve <a href="/deepseek_ai/">DeepSeek</a> V3/R1 models with large-scale expert parallelism and prefill-decode disaggregation on 96 GPUs.
It nearly matches the throughput reported by the official DeepSeek blog, achieving 52.3K input
zhyncs (@zhyncs42) 's Twitter Profile Photo

MLSys 2025 is coming up! Want to meet the developers behind FlashInfer, XGrammar, and SGLang LMSYS Org in person? Join us for the Happy Hour on May 12—we’d love to see you there! lu.ma/dl99yjoe

SkyPilot (@skypilot_org) 's Twitter Profile Photo

E2E guide to self-host multi-node Llama 4 with one command, using SkyPilot, SGLang (LMSYS Org), and `llm` (Simon Willison). ▸ Spin up SGLang on 8x H100 (single and multi-node) ▸ Perf numbers included: token/s, TTFT, ITL ▸ Production-readiness: auth, HTTPS ▸ Tooling: easy

Yixin Dong (@yi_xin_dong) 's Twitter Profile Photo

We are hosting a happy hour with LMSYS Org at #mlsys2025! Join us for engaging talks on SGLang, the structured generation library XGrammar, and the high-performance kernel library FlashInfer. Enjoy great food, lively discussions, and connect with the community! Click to join 👉

We are hosting a happy hour with <a href="/lmsysorg/">LMSYS Org</a> at #mlsys2025! Join us for engaging talks on SGLang, the structured generation library XGrammar, and the high-performance kernel library FlashInfer. Enjoy great food, lively discussions, and connect with the community! Click to join 👉
zhyncs (@zhyncs42) 's Twitter Profile Photo

I’ll be joining my Baseten colleague Philip Kiely at the AI Engineer World’s Fair AI Engineer in San Francisco, June 3–5, to Introduce LLM serving with SGLang LMSYS Org. We’d love for you to stop by and exchange ideas in person!🤗

I’ll be joining my <a href="/basetenco/">Baseten</a> colleague <a href="/philip_kiely/">Philip Kiely</a> at the AI Engineer World’s Fair <a href="/aiDotEngineer/">AI Engineer</a> in San Francisco, June 3–5, to Introduce LLM serving with SGLang <a href="/lmsysorg/">LMSYS Org</a>. We’d love for you to stop by and exchange ideas in person!🤗
Zihao Ye (@ye_combinator) 's Twitter Profile Photo

We’re thrilled that FlashInfer won a Best Paper Award at MLSys 2025! 🎉 This wouldn’t have been possible without the community — huge thanks to LMSYS Org’s sglang for deep co-design (which is crtical for inference kernel evolution) and stress-testing over the years, and to

Novita AI (@novita_labs) 's Twitter Profile Photo

Thrilled to partner with LMSYS Org! 🚀 Our high-performance GPU cloud is powering their game-changing inference engine, revolutionizing LLM deployment with breakthrough RL framework and multi-LLM serving capabilities. Together, we're bringing lightning-fast AI inference to

Thrilled to partner with <a href="/lmsysorg/">LMSYS Org</a>! 🚀

Our high-performance GPU cloud is powering their game-changing inference engine, revolutionizing LLM deployment with breakthrough RL framework and multi-LLM serving capabilities.

Together, we're bringing lightning-fast AI inference to
Benjamin F Spector (@bfspector) 's Twitter Profile Photo

(1/5) We’ve never enjoyed watching people chop Llamas into tiny pieces. So, we’re excited to be releasing our Low-Latency-Llama Megakernel! We run the whole forward pass in single kernel. Megakernels are faster & more humane. Here’s how to treat your Llamas ethically: (Joint

(1/5) We’ve never enjoyed watching people chop Llamas into tiny pieces.

So, we’re excited to be releasing our Low-Latency-Llama Megakernel! We run the whole forward pass in single kernel.

Megakernels are faster &amp; more humane. Here’s how to treat your Llamas ethically:

(Joint
zhyncs (@zhyncs42) 's Twitter Profile Photo

We sincerely appreciate NVIDIA’s generous sponsorship. It is truly regrettable that the SGLang team was unable to attend in person and witness the physical hardware firsthand. Our special thanks go to InnoMatrix for their invaluable assistance in managing everything seamlessly.