Vaibhav (VB) Srivastav (@reach_vb) 's Twitter Profile
Vaibhav (VB) Srivastav

@reach_vb

GPU poor @Huggingface | F1 fan | Here for @at_sofdog’s wisdom | *opinions my own

ID: 874987512850128897

linkhttps://huggingface.co calendar_today14-06-2017 13:50:54

5,5K Tweet

15,15K Followers

220 Following

Lewis Tunstall (@_lewtun) 's Twitter Profile Photo

Anybody can now post-train Llama 3.2 Vision on their own dataset in just a few lines of code with TRL πŸš€! We've just added support for the 11B and 90B models to the SFTTrainer, so you can fine-tune your models to both see πŸ‘€ and follow your instructions ✍️ Training script

Anybody can now post-train Llama 3.2 Vision on their own dataset in just a few lines of code with TRL πŸš€!  

We've just added support for the 11B and 90B models to the SFTTrainer, so you can fine-tune your models to both see πŸ‘€ and follow your instructions ✍️

Training script
Georgi Gerganov (@ggerganov) 's Twitter Profile Photo

Let's bring llama.cpp to the clouds! You can now run llama.cpp-powered inference endpoints through Hugging Face with just a few clicks. Simply select a GGUF model, pick your cloud provider (AWS, Azure, GCP), a suitable node GPU/CPU and you are good to go. For more info, check

Vaibhav (VB) Srivastav (@reach_vb) 's Twitter Profile Photo

🚨 Nvidia NeMo 10x faster & 4.5x more cost effective and blows OpenAI Whisper out of the water! πŸ”₯ Tops the Open ASR Leaderboard with a series of CTC, RNN-T, TDT and AED models! What powers the upto 10x speed boost: > Autocasting weights to bfloat16 > CUDA Graphs w/ conditional

🚨 Nvidia NeMo 10x faster & 4.5x more cost effective and blows OpenAI Whisper out of the water! πŸ”₯

Tops the Open ASR Leaderboard with a series of CTC, RNN-T, TDT and AED models!

What powers the upto 10x speed boost:
> Autocasting weights to bfloat16
> CUDA Graphs w/ conditional
Vaibhav (VB) Srivastav (@reach_vb) 's Twitter Profile Photo

Open Source AI was off the charts last week: Nvidia released Nemotron 51B - 220% faster and can handle 400% more workload than L3.1 70B & permissively licensed Meta dropped Llama 3.2 - Llama Vision 90B & 11B and tiny llamas (3B & 1B) for on-device usage, multilingual & with