Bye Qwen3-235B-A22B, hello Qwen3-235B-A22B-2507!
After talking with the community and thinking it through, we decided to stop using hybrid thinking mode. Instead, we’ll train Instruct and Thinking models separately so we can get the best quality possible. Today, we’re releasing
today we're releasing a new small model (0.5B) for detecting problems with tool usage in agents, trained on 50M tokens from publicly available MCP server tools
it's great at picking up on tool accuracy issues and outperforms larger models
🚀 We’re excited to introduce Qwen3-235B-A22B-Thinking-2507 — our most advanced reasoning model yet!
Over the past 3 months, we’ve significantly scaled and enhanced the thinking capability of Qwen3, achieving:
✅ Improved performance in logical reasoning, math, science & coding
🚀 Qwen3-30B-A3B-Thinking-2507, a medium-size model that can think!
• Nice performance on reasoning tasks, including math, science, code & beyond
• Good at tool use, competitive with larger models
• Native support of 256K-token context, extendable to 1M
Qwen Chat: Go to
🚀We're expanding the Tencent Hunyuan open-source LLM ecosystem with four compact models (0.5B, 1.8B, 4B, 7B)! Designed for low-power scenarios like consumer-grade GPUs, smart vehicles, smart home devices, mobile phones, and PCs, these models support cost-effective fine-tuning
Introducing Eleven Music. The highest quality AI music model.
- Complete control over genre, style, and structure
- Multi-lingual, including English, Spanish, German, Japanese and more
- Edit the sound and lyrics of individual sections or the whole song
Today we release the APIs of our Flash series, which support Qwen3-Coder and Qwen3-2507 now. Both APIs support the context length of 1M tokens. They are fast and accurate, and they are cost-effectve as well. Feel free to take a try!
Qwen3-Coder-Flash
Model Card:
We are releasing NuMarkdown-8B-Thinking, an open-source (MIT License) reasoning OCR model 🧠✨📄
NuMarkdown-8B-Thinking is apparently the first (!) reasoning VLM specialized in converting PDFs/Scans/Spreadsheets into Markdown files (typically used for RAG applications).
It
🚀 Qwen3-30B-A3B-2507 and Qwen3-235B-A22B-2507 now support ultra-long context—up to 1 million tokens!
🔧 Powered by:
• Dual Chunk Attention (DCA) – A length extrapolation method that splits long sequences into manageable chunks while preserving global coherence.
•
💡 You get 2,000 free Qwen Code runs every day!
Run this one simple command:
npx @qwen-code/qwen-code@latest
Hit Enter, and that’s it!
🚀 Now with Qwen OAuth support — super easy to use.
Try it now and supercharge your vibe code! 💻⚡
Github:github.com/QwenLM/qwen-co…
Huge computer science result:
A Tsinghua professor JUST discovered the fastest shortest path algorithm for graphs in 40yrs.
This improves on Turing award winner Tarjan’s O(m + nlogn) with Dijkstra’s, something every Computer Science student learns in college.
Mistral Document AI is now in Microsoft Azure AI Foundry!
Ever struggled to extract useful info from PDFs, scans, or complex files? Our AI understands tables, charts, and even multilingual text — turning messy documents into clean, actionable data.
Thank you, Microsoft. We’re