Image created with Flux Pro v1.1 Ultra. Image prompt: Ornate showgirl glamour in orange-and-teal tones, glittering globe centerpiece studded with jewels, stylized text “International” glowing across the curved globe’s surface in marquee lights; spotlit, dramatic contrast, vintage grain, cinematic, high-detail
Christopher Mims 🤌 on X: “The AI infrastructure build-out is so gigantic that in the past 6 months, it contributed more to the growth of the U.S. economy than /all of consumer spending/ The ‘magnificent 7’ spent more than $100 billion on data centers and the like in the past three months *alone* 1/🧵 https://t.co/sHMK1zI0sP” / X
https://x.com/mims/status/1951256592642441239
OpenAI / America is still ahead in the race”” -> no There is no western open-source model that beats or ties the best chinese open-source models.”” / X https://x.com/scaling01/status/1952900225120780705
Did yesterday’s release shift the needle in the open vs. closed debate? Today in @ReedAlbergotti’s newsletter https://x.com/fdaudens/status/1953147586312872057
I signed this because, despite worrying about misuse of open models more than most, I would like that to be the bottleneck rather than “”is it beneficial to big companies commercially/reputationally etc.”” There are many benefits to the US investing here. https://x.com/Miles_Brundage/status/1952400404668657966
RT @natolambert: America needs to take open models more seriously. This summer the early lead in open model adoption of the US via Llama ha…”” / X https://x.com/ethanCaballero/status/1952459460703834392
The relative failure of Llama 4 turned out to be very consequential to the AI landscape. It led to the shifting the locus of open weights development to China, a move towards closed models as companies running local Llama couldn’t continue to upgrade, & big talent wars in the US.”” / X https://x.com/emollick/status/1951433537485500476
The US now likely has the leading open weights models (or close to it)… … but the real question is whether this is a one-off situation from OpenAI, in which case the lead will evaporate quickly as others catch up. But also unclear what their incentives are to keep updating.”” / X https://x.com/emollick/status/1952836130958917894
Why open-source AI became an American national priority | VentureBeat
China’s ByteDance just released an LLM-based agent for general purpose software engineering tasks. Trae Agent comes with an interactive CLI that can execute complex workflows using simple English prompts. It works with OpenAI and Anthropic API. 100% opensource. https://x.com/Saboo_Shubham_/status/1942047679758151783
ByteDance dropped SeedProver. This model scored 331/657 on PutnamBench (nearly 4× better than the previous state of the art) and 201/657 under lightweight inference (pass@64‑256 equivalent). Its reported performance surpasses DeepMind’s AlphaGeometry2 and achieves 100% on https://x.com/cgeorgiaw/status/1952301113446699347
honestly scared about the power and scale of ai technologies that’ll be used in the upcoming 2028 presidential election. it could be a civilizational turning point. we aren’t ready. we should probably start preparing, or at least talking about how we could prepare.”” / X https://x.com/DavidSHolz/status/1952541453491867792
America needs to take open models more seriously. This summer the early lead in open model adoption of the US via Llama has been overtaken by Chinese models. With The American Truly Open Models (ATOM) Project we’re looking to build support and express the urgency of this issue. https://x.com/natolambert/status/1952370970762871102
very excited by the ATOM project”” / X https://x.com/finbarrtimbers/status/1952401883391520794
In partnership with the Government Services Administration, we are providing ChatGPT to the entire U.S. federal workforce for essentially no cost for the next year. https://x.com/gdb/status/1953120865115074805
OpenAI for the U.S. government:”” / X https://x.com/gdb/status/1952756538399228091
Providing ChatGPT to the entire U.S. federal workforce | OpenAI https://openai.com/index/providing-chatgpt-to-the-entire-us-federal-workforce/
The giant question is: now that The Crowd in government has access to AI tools (which, given representative surveys, many were already using) how are they going to be used to make things better, not worse? Where are Leadership & The Lab inside agencies? https://x.com/emollick/status/1953118449611272575
we are providing ChatGPT access to the entire federal workforce! (for $1 a year per agency) https://x.com/sama/status/1953103336044990779
8.6% of the world’s population uses ChatGPT weekly…”” / X https://x.com/emollick/status/1952389693502370198
Swedish Prime Minister is using AI models “”quite often”” at his job. He says he uses it get a “”second opinion”” and asks questions such as “”what have others done?”” At the moment he is not uploading any documents. IMO, when these models are capable of giving seemingly better https://x.com/rohanpaul_ai/status/1952025736111366590
Introducing Stargate Norway | OpenAI https://openai.com/index/introducing-stargate-norway/
Jensen Huang congratulates OpenAI on Stargate Norway. It will run on GB300 Superchips, scaling to hundreds of thousands of GPUs – purpose-built for training, reasoning, and real-time inference. “Just as electricity and the internet became foundational to modern life, AI will https://x.com/vitrupo/status/1950828090260955165
🚀We’re expanding the Tencent Hunyuan open-source LLM ecosystem with four compact models (0.5B, 1.8B, 4B, 7B)! Designed for low-power scenarios like consumer-grade GPUs, smart vehicles, smart home devices, mobile phones, and PCs, these models support cost-effective fine-tuning https://x.com/TencentHunyuan/status/1952262079051940322
RT @ori_press: We just benchmarked Qwen 3 Coder and GLM 4.5 on AlgoTune, and they manage to beat Claude Opus 4! We’re excited to see if the…”” / X https://x.com/OfirPress/status/1952470237947085146
ByteDance’s SeedProver scores 331/657 on PutnamBench, almost 4 times the previous SOTA. More impressively, it gets 201/657 under the *light* inference setting, ie equivalent to pass@64-256. DeepSeek-Prover-V2 is just 3 months old… Things go fast now. https://x.com/teortaxesTex/status/1951875052967739787
rule number one: never distill from DeepSeek https://x.com/jxmnop/status/1953163073612562851
wanted to share this Kimi-K2 follow-up because it’s just… so Kimi. Q: It is said that “”great things are best seen at a distance””. Do the Chinese see their country in this world-historical way at all, or do they, like Americans, primarily perceive it from a pedestrian’s point of https://x.com/teortaxesTex/status/1953280294758310310
Announcing Stargate Norway: https://x.com/gdb/status/1950785240122974412
DeepSeek-R1: 2.66 million H800 hours GPT-OSS-120B: 2.1 million H100 hours https://x.com/scaling01/status/1952784655838564376
🚀 Qwen3-30B-A3B-2507 and Qwen3-235B-A22B-2507 now support ultra-long context—up to 1 million tokens! 🔧 Powered by: • Dual Chunk Attention (DCA) – A length extrapolation method that splits long sequences into manageable chunks while preserving global coherence. • https://x.com/Alibaba_Qwen/status/1953760230141309354
Qwen3-Coder is now available on Cerebras, 17x faster than on GPU providers. And it’s completely free. Try it out directly in your developer flow, or signup for our virtual hackathon tomorrow. It’s a $5,000 prize 🙂 @CerebrasSystems @cline https://x.com/SarahChieng/status/1951453803905163693
Small but mighty! Qwen3-Coder-Flash and GLM-4.5-Air are now on @FireworksAI_HQ Despite being smaller and faster, Qwen3 Coder Flash 30B and GLM 4.5-Air achieve almost the same quality as their larger counterparts on tool use benchmarks. The secret of good model behavior is in https://x.com/dzhulgakov/status/1952049826067050735
🚀 Meet Qwen-Image — a 20B MMDiT model for next-gen text-to-image generation. Especially strong at creating stunning graphic posters with native text. Now open-source. 🔍 Key Highlights: 🔹 SOTA text rendering — rivals GPT-4o in English, best-in-class for Chinese 🔹 In-pixel https://x.com/Alibaba_Qwen/status/1952398250121756992
💡 You get 2,000 free Qwen Code runs every day! Run this one simple command: npx @qwen-code/qwen-code@latest Hit Enter, and that’s it! 🚀 Now with Qwen OAuth support — super easy to use. Try it now and supercharge your vibe code! 💻⚡ Github: https://x.com/Alibaba_Qwen/status/1953835877555151134
Just included example scripts for aligning models using GSPO (including VLM example) 🙆♂️🙆♂️ GSPO is the latest RL alignment algo by @Alibaba_Qwen and it’s already supported in the latest TRL v0.20 release. Super-easy-to-get-started example scripts below, GO run them! 👩💻👩💻 https://x.com/SergioPaniego/status/1952305247411691871
Qwen-Image demo on Hugging Face getting absolutely hammered right now 😀 https://x.com/victormustar/status/1952416615351366033
Qwen-Image: Crafting with Native Text Rendering | Qwen https://qwenlm.github.io/blog/qwen-image/
RT @Alibaba_Qwen: 🚀 Introducing Qwen3-4B-Instruct-2507 & Qwen3-4B-Thinking-2507 — smarter, sharper, and 256K-ready! 🔹 Instruct: Boosted ge…”” / X https://x.com/NandoDF/status/1953223478087143640
RT @Alibaba_Qwen: 🚀 Meet Qwen-Image — a 20B MMDiT model for next-gen text-to-image generation. Especially strong at creating stunning graph…”” / X https://x.com/mervenoyann/status/1952455331205841261
So, I did some coding this week… – Qwen3 Coder Flash (30B-A3B) – Mixture-of-Experts setup with 128 experts, 8 active per token – In pure PyTorch (optimized for human readability) – in a standalone Jupyter notebook – Runs on a single A100 https://x.com/rasbt/status/1951635208375034191
Today we release the APIs of our Flash series, which support Qwen3-Coder and Qwen3-2507 now. Both APIs support the context length of 1M tokens. They are fast and accurate, and they are cost-effectve as well. Feel free to take a try! Qwen3-Coder-Flash Model Card:”” / X https://x.com/Alibaba_Qwen/status/1952767585596145773
@ostrisai The VAE is a fine-tune from the Wan 2.1 VAE for image generation, which is super cool and shows how open source foster collaboration, even between rival labs”” / X https://x.com/multimodalart/status/1952409238413684901
China’s cranking out humanoids. NOETIX delivered 105 robots in July alone. https://x.com/TheHumanoidHub/status/1951330695986729435




