Image created with GPT Image 1. Image prompt: A silk-robed Black merchant prince with gold calligraphy lining and a jade-embellished turban stands at the Met Gala flanked by mirrored camels bearing QR codes and silks, referencing ancient trade routes and modern e-commerce empires; shot in golden hour light with East-meets-West regal fusion.
Qwen3 benchmark results 235B is a BEAST placing 3rd in the overall and with the best generalization among all tested models all of the Qwen3 models have very low or perfect percentages of invalid moves which means good instruction following 235B MoE > 32B > 14B > 30B MoE > 8B https://x.com/scaling01/status/1918031153312731536
2/ China’s Alibaba just released Qwen 3 with support for MCP and 119 languages. It matches the performance of DeepSeek-R1, OpenAI o1, o3-mini, and Grok-3. @Saboo_Shubham_ Plus, AI Agents with Qwen3 can now think deeper with hybrid reasoning modes. https://x.com/AtomSilverman/status/1918424770749874668
China’s Alibaba just released Qwen 3 with support for MCP and 119 languages. It matches the performance of DeepSeek-R1, OpenAI o1, o3-mini, and Grok-3. Plus, AI Agents with Qwen3 can now think deeper with hybrid reasoning modes. https://x.com/Saboo_Shubham_/status/1916972515077066922
We’re launching Computer Use in smolagents! 🥳 -> As vision models become more capable, they become able to power complex agentic workflows. Especially Qwen-VL models, that support built-in grounding, i.e. ability to locate any element in an image by its coordinates, thus to https://x.com/AymericRoucher/status/1919783847597670780
Is Qwen3-235B the new budget-friendly coding champ in Cline? Early user feedback is rolling in — it’s promising, but not perfect. Here’s what we’re hearing from the Cline community: 🧵” / X https://x.com/cline/status/1917708041857949983
Alibaba’s Qwen team released Qwen3 family with 2 MoE models and 6 dense models —Models range from 600M to 235B params —Flagship version rivals OpenAI o1 & DeepSeek-R1 —Hybrid “thinking” in all —Boosted coding + agent performance —119 languages supported https://x.com/adcock_brett/status/1919060402417119375
👏🏻 Excited to see Qwen3-235B-A22B’s impressive performance on LiveCodeBench! This positions Qwen3 as the top open model for competitive-level code generation, matching the performance of o4-mini (low). https://x.com/huybery/status/1919418019517776024
Pretty fucking incredible week so far: > Qwen3 – MoE (235B, 30B) + Dense (32, 14, 8, 4, 0.6B) > Xiaomi – MiMo 7B dense > Kyutai – Helium 2B dense > DeepSeek – Prover V2 671B MoE > Qwen2.5 Omni 3B > Microsoft – Phi4 14B Reasoning, Mini (3.8B) & Plus > JetBrains- Mellum 4B Dense” / X https://x.com/reach_vb/status/1917938596465750476
The community votes are in for Qwen3-235B-A22B 🥁 The latest open-source Qwen3 is now on the Arena Top 10 🏆 Congrats to @alibaba_qwen on this achievement! 👏 Highlights: 💠 For Chat: Qwen3-235B-A22B ranks #10, tied with o1 💠 Strong in Coding at #4 and Math #1 💠 For WebDev: https://x.com/lmarena_ai/status/1919448953042706759
Alibaba unveils Qwen3, a family of ‘hybrid’ AI reasoning models | TechCrunch https://techcrunch.com/2025/04/28/alibaba-unveils-qwen-3-a-family-of-hybrid-ai-reasoning-models/
@Alibaba_Qwen Do you plan to make a Qwen 3 Coder in the future with FIM capabilities similar to Qwen 2.5 Coder?” / X https://x.com/ggerganov/status/1918373399891513571
Is this the Smol Models Festival? – Phi-4 just dropped a new reasoning model – Qwen2.5 Omni now has a 3B version – And OLMo-2 1B version https://x.com/fdaudens/status/1917961029675347973
Qwen 3 235B now on @togethercompute API! Qwen 3 is a reasoning model that has a non-reasoning instruct mode with allowance for setting a thinking budget. It’s efficient ($0.20/M input & $0.60/M output on our throughput optimized endpoint) and fantastic on a variety of” / X https://x.com/vipulved/status/1917777842466889873
A ton of impactful models and datasets in open AI past week, let’s summarize the best 🤩 link to all are on the next one ⤵️ 💬 @Alibaba_Qwen made it rain! They released Qwen3: new dense and MoE models ranging from 0.6B to 235B 🤯 as well as Qwen2.5-Omni, any-to-any model in 3B https://x.com/mervenoyann/status/1919784802099540446
We will release the quantized models of Qwen3 to you in the following days. Today we release the AWQ and GGUFs of Qwen3-14B and Qwen3-32B, which enables using the models with limited GPU memory. Qwen3-32B-AWQ: https://x.com/Alibaba_Qwen/status/1918353505074725363
Introducing Qwen3! We release and open-weight Qwen3, our latest large language models, including 2 MoE models and 6 dense models, ranging from 0.6B to 235B. Our flagship model, Qwen3-235B-A22B, achieves competitive results in benchmark evaluations of coding, math, general https://x.com/Alibaba_Qwen/status/1916962087676612998



