vLLM (@vllm_project) 's Twitter Profile
vLLM

@vllm_project

A high-throughput and memory-efficient inference and serving engine for LLMs. Join slack.vllm.ai to discuss together with the community!

ID: 1774187564276289536

linkhttps://github.com/vllm-project/vllm calendar_today30-03-2024 21:31:01

327 Tweet

12,12K Followers

15 Following

vLLM (@vllm_project) 's Twitter Profile Photo

⬆️ uv pip install -U vLLM The latest release features 649 commits from 215 contributors. vLLM is now ready for NVIDIA Blackwell with the latest PyTorch 2.7 upgrade. Huge thanks to NVIDIA AI Developer and Zihao Ye for the CUTLASS and FlashInfer kernels!

⬆️ uv pip install -U vLLM
The latest release features 649 commits from 215 contributors. vLLM is now ready for <a href="/nvidia/">NVIDIA</a> Blackwell with the latest <a href="/PyTorch/">PyTorch</a> 2.7 upgrade. Huge thanks to <a href="/NVIDIAAIDev/">NVIDIA AI Developer</a> and <a href="/ye_combinator/">Zihao Ye</a> for the CUTLASS and FlashInfer kernels!