Axel πŸ‡¬πŸ‡§ Dev | Software Engineer | AI Claude GPT (@axelgarciak) 's Twitter Profile
Axel πŸ‡¬πŸ‡§ Dev | Software Engineer | AI Claude GPT

@axelgarciak

πŸ‘¨β€πŸ’» Software Engineer πŸ’Ύ Software retro minimalist πŸ€– AI tinkerer πŸ—οΈ Building tech communities πŸ‡¬πŸ‡§πŸ‡©πŸ‡ͺπŸ‡»πŸ‡ͺ

ID: 1506396121693884419

linkhttp://axelgarciak.com/bio calendar_today22-03-2022 22:23:34

8,8K Tweet

4,4K Followers

652 Following

Axel πŸ‡¬πŸ‡§ Dev | Software Engineer | AI Claude GPT (@axelgarciak) 's Twitter Profile Photo

Qwen3-Max-Preview available on anycoder. πŸ’» A shame it's not an open source model but shout-out to the Qwen team for releasing so many great open source models in the past.

AiBattle (@aibattle_) 's Twitter Profile Photo

New Qwen-3-Next-80B-A3B model incoming "Built on this architecture, we trained and open-sourced Qwen3-Next-80B-A3B β€” 80B total parameters, only 3B active β€” achieving extreme sparsity and efficiency. Despite its ultra-efficiency, it outperforms Qwen3-32B on downstream tasks β€”

New Qwen-3-Next-80B-A3B model incoming

"Built on this architecture, we trained and open-sourced Qwen3-Next-80B-A3B β€” 80B total parameters, only 3B active β€” achieving extreme sparsity and efficiency.

Despite its ultra-efficiency, it outperforms Qwen3-32B on downstream tasks β€”
AK (@_akhaliq) 's Twitter Profile Photo

ERNIE-4.5-21B-A3B-Thinking a text MoE post-trained model, with 21B total parameters and 3B activated parameters for each token. Significantly improved performance on reasoning tasks, including logical reasoning, mathematics, science, coding, text generation, and academic

ERNIE-4.5-21B-A3B-Thinking

a text MoE post-trained model, with 21B total parameters and 3B activated parameters for each token.

Significantly improved performance on reasoning tasks, including logical reasoning, mathematics, science, coding, text generation, and academic
Axel πŸ‡¬πŸ‡§ Dev | Software Engineer | AI Claude GPT (@axelgarciak) 's Twitter Profile Photo

I like the new wave of models with a medium to high total parameter count and a low active parameter counts like Qwen3 and ERNIE-4.5. These models are perfect to run in systems like AMD Ryzen Strix Halo with 96GB or 128GB of RAM.

Axel πŸ‡¬πŸ‡§ Dev | Software Engineer | AI Claude GPT (@axelgarciak) 's Twitter Profile Photo

Goodbye to 2025 πŸ‘‹ Happy New Year 2026! Small LLMs highlights in 2025: - Qwen3-4B - LFM2–8B-A1B / LFM2-2.6B-Exp - granite-4.0-h-tiny (7B A1B) - gemma-3n-E4B / gemma-3-4b-it - SmolLM3-3B - Ministral-3-3B They run at decent speeds on the CPU! Looking forward to 2026's smol LLMs!

Ramin Hasani (@ramin_m_h) 's Twitter Profile Photo

For us at Liquid, the goal was always to fundamentally bring efficiency to the point that you can take it for granted: models that run light, fast, on any processor. The art is can we obtain this efficiency for small models with the quality and reliability that can power real

For us at Liquid, the goal was always to fundamentally bring efficiency to the point that you can take it for granted: models that run light, fast, on any processor. The art is can we obtain this efficiency for small models with the quality and reliability that can power real