Image created with Flux Pro v1.1 Ultra. Image prompt: Ornate showgirl glamour in orange-and-teal tones, dazzling feathered crown prop sparkling midstage, stylized text “Qwen” glowing in pearl-studded neon script across the crown; spotlit, dramatic contrast, vintage grain, cinematic, high-detail

RT @ori_press: We just benchmarked Qwen 3 Coder and GLM 4.5 on AlgoTune, and they manage to beat Claude Opus 4! We’re excited to see if the…”” / X https://x.com/OfirPress/status/1952470237947085146

🚀 Qwen3-30B-A3B-2507 and Qwen3-235B-A22B-2507 now support ultra-long context—up to 1 million tokens! 🔧 Powered by: • Dual Chunk Attention (DCA) – A length extrapolation method that splits long sequences into manageable chunks while preserving global coherence. • https://x.com/Alibaba_Qwen/status/1953760230141309354

Qwen3-Coder is now available on Cerebras, 17x faster than on GPU providers. And it’s completely free. Try it out directly in your developer flow, or signup for our virtual hackathon tomorrow. It’s a $5,000 prize 🙂 @CerebrasSystems @cline https://x.com/SarahChieng/status/1951453803905163693

Small but mighty! Qwen3-Coder-Flash and GLM-4.5-Air are now on @FireworksAI_HQ Despite being smaller and faster, Qwen3 Coder Flash 30B and GLM 4.5-Air achieve almost the same quality as their larger counterparts on tool use benchmarks. The secret of good model behavior is in https://x.com/dzhulgakov/status/1952049826067050735

🚀 Meet Qwen-Image — a 20B MMDiT model for next-gen text-to-image generation. Especially strong at creating stunning graphic posters with native text. Now open-source. 🔍 Key Highlights: 🔹 SOTA text rendering — rivals GPT-4o in English, best-in-class for Chinese 🔹 In-pixel https://x.com/Alibaba_Qwen/status/1952398250121756992

💡 You get 2,000 free Qwen Code runs every day! Run this one simple command: npx @​qwen-code/qwen-code@latest Hit Enter, and that’s it! 🚀 Now with Qwen OAuth support — super easy to use. Try it now and supercharge your vibe code! 💻⚡ Github: https://x.com/Alibaba_Qwen/status/1953835877555151134

Just included example scripts for aligning models using GSPO (including VLM example) 🙆‍♂️🙆‍♂️ GSPO is the latest RL alignment algo by @Alibaba_Qwen and it’s already supported in the latest TRL v0.20 release. Super-easy-to-get-started example scripts below, GO run them! 👩‍💻👩‍💻 https://x.com/SergioPaniego/status/1952305247411691871

Qwen-Image demo on Hugging Face getting absolutely hammered right now 😀 https://x.com/victormustar/status/1952416615351366033

Qwen-Image: Crafting with Native Text Rendering | Qwen https://qwenlm.github.io/blog/qwen-image/

RT @Alibaba_Qwen: 🚀 Introducing Qwen3-4B-Instruct-2507 & Qwen3-4B-Thinking-2507 — smarter, sharper, and 256K-ready! 🔹 Instruct: Boosted ge…”” / X https://x.com/NandoDF/status/1953223478087143640

RT @Alibaba_Qwen: 🚀 Meet Qwen-Image — a 20B MMDiT model for next-gen text-to-image generation. Especially strong at creating stunning graph…”” / X https://x.com/mervenoyann/status/1952455331205841261

So, I did some coding this week… – Qwen3 Coder Flash (30B-A3B) – Mixture-of-Experts setup with 128 experts, 8 active per token – In pure PyTorch (optimized for human readability) – in a standalone Jupyter notebook – Runs on a single A100 https://x.com/rasbt/status/1951635208375034191

Today we release the APIs of our Flash series, which support Qwen3-Coder and Qwen3-2507 now. Both APIs support the context length of 1M tokens. They are fast and accurate, and they are cost-effectve as well. Feel free to take a try! Qwen3-Coder-Flash Model Card:”” / X https://x.com/Alibaba_Qwen/status/1952767585596145773

@ostrisai The VAE is a fine-tune from the Wan 2.1 VAE for image generation, which is super cool and shows how open source foster collaboration, even between rival labs”” / X https://x.com/multimodalart/status/1952409238413684901

Trending

Discover more from Ethan B. Holland

Subscribe now to keep reading and get access to the full archive.

Continue reading