Jingwei Zuo (@jingweizuo) 's Twitter Profile
Jingwei Zuo

@jingweizuo

Lead Researcher @tiiuae, Falcon LLM team huggingface.co/tiiuae

ID: 1065296158669582339

linkhttps://jingweizuo.com calendar_today21-11-2018 17:29:31

46 Tweet

61 Takipçi

77 Takip Edilen

Hongyu Wang (@realhongyu_wang) 's Twitter Profile Photo

We just released the fine-tuning code and fine-tuned models of BitVLA in Huggingface🔥🔥 Enjoy these hyper-efficient 1-bit VLA models! Code: github.com/ustcwhy/BitVLA Models: huggingface.co/collections/ho…

Jingwei Zuo (@jingweizuo) 's Twitter Profile Photo

We’re organizing the E2LM competition at #NeurIPS2025, focused on early-stage training evaluations of Large Language Models. Registration is now open — join us and help revolutionize how we evaluate LLMs! 🚀

Technology Innovation Institute (@tiiuae) 's Twitter Profile Photo

🚀 Exciting news! Falcon-H1 & Falcon-E are now on Oumi — the open-source platform for training, fine-tuning (SFT, LoRA, QLoRA), and deploying LLMs anywhere: laptops, cloud, or clusters. Start building: github.com/oumi-ai/oumi/t… #FalconH1 #FalconE #OpenSourceAI #LLM

🚀 Exciting news! Falcon-H1 & Falcon-E are now on Oumi — the open-source platform for training, fine-tuning (SFT, LoRA, QLoRA), and deploying LLMs anywhere: laptops, cloud, or clusters.

Start building: github.com/oumi-ai/oumi/t…

#FalconH1 #FalconE #OpenSourceAI #LLM
Awni Hannun (@awnihannun) 's Twitter Profile Photo

Latest mlx-lm is out! pip install -U mlx-lm Bunch of new models: - SmolLM3 (Hugging Face) - Ernie family (Baidu) - BitNet (Microsoft) - Falcon-E (TII) - Text-only Gemma3n (Google) - MiniCPM4 (OpenBMB) - AFM (Apple) +Performance improvements for DWQ, dynamic quantization, and

Latest mlx-lm is out!

pip install -U mlx-lm

Bunch of new models:
- SmolLM3 (Hugging Face)
- Ernie family (Baidu)
- BitNet (Microsoft)
- Falcon-E (TII)
- Text-only Gemma3n (Google)
- MiniCPM4 (OpenBMB)
- AFM (Apple)

+Performance improvements for DWQ, dynamic quantization, and
younes (@younesbelkada) 's Twitter Profile Photo

Excited to have contributed into Falcon-E (Bitnet) integration with Prince Canuma Awni Hannun in mlx-lm Falcon-E now fully supported in mlx-lm - as simple as `mlx_lm.generate --model tiiuae/Falcon-E-1B-Instruct --prompt "Implement bubble sort" --max-tokens 100 --temp 0.1` 🚀

Excited to have contributed into Falcon-E (Bitnet) integration with <a href="/Prince_Canuma/">Prince Canuma</a> <a href="/awnihannun/">Awni Hannun</a>  in mlx-lm
Falcon-E now fully supported in mlx-lm - as simple as `mlx_lm.generate --model tiiuae/Falcon-E-1B-Instruct --prompt "Implement bubble sort" --max-tokens 100 --temp 0.1` 🚀
Technology Innovation Institute (@tiiuae) 's Twitter Profile Photo

Falcon-H1 now runs natively on your device via llama.cpp—0.5B to 34B models, no server needed. Fast inference, long context, multilingual, tool-ready. Build, test, and go beyond. #FalconH1 #LocalLLM #AIOnDevice #EdgeAI #OpenSourceAI

Rosinality (@rosinality) 's Twitter Profile Photo

Falcon-H1: A Family of Hybrid-Head Language Models Redefining Efficiency and Performance Falcon's side-by-side attention-SSM hybrid model. Very detailed, from tokenizers to data preparation and optimization strategies.

Falcon-H1: A Family of Hybrid-Head Language Models Redefining Efficiency and Performance

Falcon's side-by-side attention-SSM hybrid model. Very detailed, from tokenizers to data preparation and optimization strategies.
Teortaxes▶️ (DeepSeek 推特🐋铁粉 2023 – ∞) (@teortaxestex) 's Twitter Profile Photo

Falcon-H1 is a very dense research paper exploring the space of hybrid attention designs and tuning *every* hyperparameter there is. It's more interesting than models themselves. If you were intrigued by that «AlphaGo move» slop, this is the real thing.

Falcon-H1 is a very dense research paper exploring the space of hybrid attention designs and tuning *every* hyperparameter there is. It's more interesting than models themselves. If you were intrigued by that «AlphaGo move» slop, this is the real thing.
Jingwei Zuo (@jingweizuo) 's Twitter Profile Photo

It’s concerning how recent model releases flex on a handful of benchmarks - but miss the bigger picture: world knowledge, nuance, common sense. Are we still building foundation models, or just performance models? VIBE CHECK: that’s all you need.

Si-ze Zheng (@deeplyignorant) 's Twitter Profile Photo

🎉 Excited to share: We’ve open-sourced Triton-distributed MegaKernel! A fresh, powerful take on MegaKernel for LLMs—built entirely on our Triton-distributed framework. github.com/ByteDance-Seed… Why it’s awesome? 🧩 Super programmable ⚡ Blazing performance 📊 Rock-solid precision

🎉 Excited to share: We’ve open-sourced Triton-distributed MegaKernel! A fresh, powerful take on MegaKernel for LLMs—built entirely on our Triton-distributed framework.
github.com/ByteDance-Seed…

Why it’s awesome?
🧩 Super programmable
⚡ Blazing performance
📊 Rock-solid precision
vLLM (@vllm_project) 's Twitter Profile Photo

🚀 Amazing community project! vLLM CLI — a command-line tool for serving LLMs with vLLM: ✅ Interactive menu-driven UI & scripting-friendly CLI ✅ Local + HuggingFace Hub model management ✅ Config profiles for perf/memory tuning ✅ Real-time server & GPU monitoring ✅ Error

🚀 Amazing community project!

vLLM CLI — a command-line tool for serving LLMs with vLLM:
✅ Interactive menu-driven UI &amp; scripting-friendly CLI
✅ Local + HuggingFace Hub model management
✅ Config profiles for perf/memory tuning
✅ Real-time server &amp; GPU monitoring
✅ Error
Ai2 (@allen_ai) 's Twitter Profile Photo

📢 New paper from Ai2: Signal & Noise asks a simple question—can language model benchmarks detect a true difference in model performance? 🧵

📢 New paper from Ai2: Signal &amp; Noise asks a simple question—can language model benchmarks detect a true difference in model performance? 🧵