Lysandre (@lysandrejik) 's Twitter Profile
Lysandre

@lysandrejik

Chief Open-Source Officer (COSO) at Hugging Face

ID: 1105862126432894976

calendar_today13-03-2019 16:04:11

1,1K Tweet

8,8K Followers

603 Following

vLLM (@vllm_project) 's Twitter Profile Photo

The Hugging Face Transformers โ†”๏ธ vLLM integration just leveled up: Vision-Language Models are now supported out of the box! If the model is integrated into Transformers, you can now run it directly with vLLM. github.com/vllm-project/vโ€ฆ Great work Raushan Turganbay ๐Ÿ‘

The <a href="/huggingface/">Hugging Face</a> Transformers โ†”๏ธ <a href="/vllm_project/">vLLM</a> integration just leveled up: Vision-Language Models are now supported out of the box!  

If the model is integrated into Transformers, you can now run it directly with vLLM.

github.com/vllm-project/vโ€ฆ

Great work <a href="/RTurganbay/">Raushan Turganbay</a> ๐Ÿ‘
AK (@_akhaliq) 's Twitter Profile Photo

Hugging Face just dropped Trending Papers partnered with AI at Meta and Papers with Code to build a successor to papers with code (which was sunsetted yesterday) HF papers provides a new section for the community to follow trending Papers, linked to their code implementations

Hugging Face just dropped Trending Papers

partnered with <a href="/AIatMeta/">AI at Meta</a> and <a href="/paperswithcode/">Papers with Code</a> to build a successor to papers with code  (which was sunsetted yesterday)  

HF papers provides a new section for the community to follow trending Papers, linked to their code implementations
Arthur Zucker (@art_zucker) 's Twitter Profile Photo

With the latest release, I want to make sure I get this message to the community: we are listening! Hugging Face we are very ambitious and we want `transformers` to accelerate the ecosystem and enable all hardwares / platforms! Let's build AGI together ๐Ÿซฃ Unbloat and Enable!

With the latest release, I want to make sure I get this message to the community: we are listening! 

<a href="/huggingface/">Hugging Face</a> we are very ambitious and we want `transformers` to accelerate the ecosystem and enable all hardwares / platforms! 
Let's build AGI together ๐Ÿซฃ
Unbloat and Enable!
๐Ÿ‘‹ Jan (@jandotai) 's Twitter Profile Photo

Hugging Face ๐Ÿค Jan You can now use Hugging Face as a remote model provider in Jan. Go to Settings -> Model Providers -> add your Hugging Face API key. Then open a new chat and pick a model from Hugging Face. Works with any model in Hugging Face in Jan.

LeRobot (@lerobothf) 's Twitter Profile Photo

๐Ÿš€ ๐„๐ฑ๐œ๐ข๐ญ๐ข๐ง๐  ๐๐ž๐ฐ๐ฌ! ๐๐ฒ๐๐ˆ ๐š๐ง๐ ๐†๐ข๐ญ๐‡๐ฎ๐› ๐‘๐ž๐ฅ๐ž๐š๐ฌ๐ž ๐Ÿ๐จ๐ซ ๐‹๐ž๐‘๐จ๐›๐จ๐ญ ๐Ÿš€ Weโ€™re thrilled to announce that we have reached a new major milestone, the first ๐จ๐Ÿ๐Ÿ๐ข๐œ๐ข๐š๐ฅ ๐๐ฒ๐๐ˆ and ๐†๐ข๐ญ๐‡๐ฎ๐› ๐‘๐ž๐ฅ๐ž๐š๐ฌ๐ž! ๐ŸŽ‰ * You can now install LeRobot with a

๐Ÿš€ ๐„๐ฑ๐œ๐ข๐ญ๐ข๐ง๐  ๐๐ž๐ฐ๐ฌ! ๐๐ฒ๐๐ˆ ๐š๐ง๐ ๐†๐ข๐ญ๐‡๐ฎ๐› ๐‘๐ž๐ฅ๐ž๐š๐ฌ๐ž ๐Ÿ๐จ๐ซ ๐‹๐ž๐‘๐จ๐›๐จ๐ญ ๐Ÿš€

Weโ€™re thrilled to announce that we have reached a new major milestone, the first ๐จ๐Ÿ๐Ÿ๐ข๐œ๐ข๐š๐ฅ ๐๐ฒ๐๐ˆ and ๐†๐ข๐ญ๐‡๐ฎ๐› ๐‘๐ž๐ฅ๐ž๐š๐ฌ๐ž! ๐ŸŽ‰

* You can now install LeRobot with a
Teknium (e/ฮป) (@teknium1) 's Twitter Profile Photo

All the details of OpenAI's new base model courtesy of HuggingFace update log. - Looks like NO base model (despite their oss model cookbook page saying it is) - 21B and 117B Total Param, 3.6B and 5.1B Active MoE Model sizes - Reasoning and Agentic capabilities - License: APACHE

Sam Altman (@sama) 's Twitter Profile Photo

gpt-oss is out! we made an open model that performs at the level of o4-mini and runs on a high-end laptop (WTF!!) (and a smaller one that runs on a phone). super proud of the team; big triumph of technology.

Xenova (@xenovacom) 's Twitter Profile Photo

HUGE: OpenAI just released GPT OSS on Hugging Face! ๐Ÿคฏ Here's what you need to know: 1. Two models: gpt-oss-20b and gpt-oss-120b, designed for powerful reasoning, agentic tasks, and more! 2. Mixture-of-experts (MoE) architecture: 21B and 117B total parameters, with 3.6B and 5.1B

HUGE: OpenAI just released GPT OSS on Hugging Face! ๐Ÿคฏ

Here's what you need to know:
1. Two models: gpt-oss-20b and gpt-oss-120b, designed for powerful reasoning, agentic tasks, and more!
2. Mixture-of-experts (MoE) architecture: 21B and 117B total parameters, with 3.6B and 5.1B
Matthew Carrigan (@carrigmat) 's Twitter Profile Photo

GPT OSS is out. It's OpenAI's first open-weights model release since GPT-2, and some of the technical innovations have huge implications. This is a thread about two of them: Learned attention sinks, and MXFP4 weights.

GPT OSS is out. It's OpenAI's first open-weights model release since GPT-2, and some of the technical innovations have huge implications. This is a thread about two of them: Learned attention sinks, and MXFP4 weights.
clem ๐Ÿค— (@clementdelangue) 's Twitter Profile Photo

When Sam Altman told me at the AI summit in Paris that they were serious about releasing open-source models & asked what would be useful, I couldnโ€™t believe it. But six months of collaboration later, here it is: Welcome to OSS-GPT on Hugging Face! It comes in two sizes, for both

When <a href="/sama/">Sam Altman</a> told me at the AI summit in Paris that they were serious about releasing open-source models &amp; asked what would be useful, I couldnโ€™t believe it. 

But six months of collaboration later, here it is: Welcome to OSS-GPT on <a href="/huggingface/">Hugging Face</a>! It comes in two sizes, for both
dylan (@dylan_ebert_) 's Twitter Profile Photo

OpenAI just released GPT-OSS: An Open Source Language Model on Hugging Face Open source meaning: ๐Ÿ’ธ Free ๐Ÿ”’ Private ๐Ÿ”ง Customizable

Mohamed (@mekkcyber) 's Twitter Profile Photo

The new GPT-OSS models are Mixture of Experts (MoEs), with 20B and 120B parameters. Since expert weights make up ~90% of the model, OpenAI decided to quantize them to 4 bits during post-training using the MXFP4 standard. Quantizing these to MXFP4 enables the larger model to

The new GPT-OSS models are Mixture of Experts (MoEs), with 20B and 120B parameters.

Since expert weights make up ~90% of the model, OpenAI decided to quantize them to 4 bits during post-training using the MXFP4 standard. 

Quantizing these to MXFP4 enables the larger model to
LLaMA Factory (@llamafactory_ai) 's Twitter Profile Photo

๐Ÿš€ LlamaFactory now supports the fine-tuning of GPT-OSS models. Check out this thread to get started in 10 minutes ๐Ÿค— github.com/hiyouga/LLaMA-โ€ฆ

๐Ÿš€ LlamaFactory now supports the fine-tuning of GPT-OSS models.

Check out this thread to get started in 10 minutes ๐Ÿค—
github.com/hiyouga/LLaMA-โ€ฆ
Matej Sirovatka (@m_sirovatka) 's Twitter Profile Photo

We have cooked some something nice with Axolotl for ๐Ÿค— accelerate v1.10. Have you ever wanted to train a large model, but couldn't setup your env? Current multi-gpu frameworks are hard to install, the configuration has like 100 options. Here comes ParallelismConfig ๐Ÿš€ 1/5 ๐Ÿงต

We have cooked some something nice with <a href="/axolotl_ai/">Axolotl</a> for ๐Ÿค— accelerate v1.10.
Have you ever wanted to train a large model, but couldn't setup your env? Current multi-gpu frameworks are hard to install, the configuration has like 100 options. Here comes ParallelismConfig ๐Ÿš€ 1/5 ๐Ÿงต
Wing Lian (caseus) (@winglian) 's Twitter Profile Photo

This is by far my favorite feature so far this year. Unlocks a lot of exciting possibilities for post-training larger models in the Hugging Face ecosystem such as improved multi-node and slurm support.