Image created with Flux Pro v1.1 Ultra. Image prompt: Ornate showgirl glamour in orange-and-teal tones, glittering globe centerpiece studded with jewels, stylized text “International” glowing across the curved globe’s surface in marquee lights; spotlit, dramatic contrast, vintage grain, cinematic, high-detail

Christopher Mims 🤌 on X: “The AI infrastructure build-out is so gigantic that in the past 6 months, it contributed more to the growth of the U.S. economy than /all of consumer spending/ The ‘magnificent 7’ spent more than $100 billion on data centers and the like in the past three months *alone* 1/🧵 https://t.co/sHMK1zI0sP” / X
https://x.com/mims/status/1951256592642441239

OpenAI / America is still ahead in the race”” -> no There is no western open-source model that beats or ties the best chinese open-source models.”” / X https://x.com/scaling01/status/1952900225120780705

Did yesterday’s release shift the needle in the open vs. closed debate? Today in @ReedAlbergotti’s newsletter https://x.com/fdaudens/status/1953147586312872057

I signed this because, despite worrying about misuse of open models more than most, I would like that to be the bottleneck rather than “”is it beneficial to big companies commercially/reputationally etc.”” There are many benefits to the US investing here. https://x.com/Miles_Brundage/status/1952400404668657966

RT @natolambert: America needs to take open models more seriously. This summer the early lead in open model adoption of the US via Llama ha…”” / X https://x.com/ethanCaballero/status/1952459460703834392

The relative failure of Llama 4 turned out to be very consequential to the AI landscape. It led to the shifting the locus of open weights development to China, a move towards closed models as companies running local Llama couldn’t continue to upgrade, & big talent wars in the US.”” / X https://x.com/emollick/status/1951433537485500476

The US now likely has the leading open weights models (or close to it)… … but the real question is whether this is a one-off situation from OpenAI, in which case the lead will evaporate quickly as others catch up. But also unclear what their incentives are to keep updating.”” / X https://x.com/emollick/status/1952836130958917894

Why open-source AI became an American national priority | VentureBeat

Why open-source AI became an American national priority

China’s ByteDance just released an LLM-based agent for general purpose software engineering tasks. Trae Agent comes with an interactive CLI that can execute complex workflows using simple English prompts. It works with OpenAI and Anthropic API. 100% opensource. https://x.com/Saboo_Shubham_/status/1942047679758151783

ByteDance dropped SeedProver. This model scored 331/657 on PutnamBench (nearly 4× better than the previous state of the art) and 201/657 under lightweight inference (pass@64‑256 equivalent). Its reported performance surpasses DeepMind’s AlphaGeometry2 and achieves 100% on https://x.com/cgeorgiaw/status/1952301113446699347

honestly scared about the power and scale of ai technologies that’ll be used in the upcoming 2028 presidential election. it could be a civilizational turning point. we aren’t ready. we should probably start preparing, or at least talking about how we could prepare.”” / X https://x.com/DavidSHolz/status/1952541453491867792

America needs to take open models more seriously. This summer the early lead in open model adoption of the US via Llama has been overtaken by Chinese models. With The American Truly Open Models (ATOM) Project we’re looking to build support and express the urgency of this issue. https://x.com/natolambert/status/1952370970762871102

very excited by the ATOM project”” / X https://x.com/finbarrtimbers/status/1952401883391520794

In partnership with the Government Services Administration, we are providing ChatGPT to the entire U.S. federal workforce for essentially no cost for the next year. https://x.com/gdb/status/1953120865115074805

OpenAI for the U.S. government:”” / X https://x.com/gdb/status/1952756538399228091

Providing ChatGPT to the entire U.S. federal workforce | OpenAI https://openai.com/index/providing-chatgpt-to-the-entire-us-federal-workforce/

The giant question is: now that The Crowd in government has access to AI tools (which, given representative surveys, many were already using) how are they going to be used to make things better, not worse? Where are Leadership & The Lab inside agencies? https://x.com/emollick/status/1953118449611272575

we are providing ChatGPT access to the entire federal workforce! (for $1 a year per agency) https://x.com/sama/status/1953103336044990779

8.6% of the world’s population uses ChatGPT weekly…”” / X https://x.com/emollick/status/1952389693502370198

Swedish Prime Minister is using AI models “”quite often”” at his job. He says he uses it get a “”second opinion”” and asks questions such as “”what have others done?”” At the moment he is not uploading any documents. IMO, when these models are capable of giving seemingly better https://x.com/rohanpaul_ai/status/1952025736111366590

Introducing Stargate Norway | OpenAI https://openai.com/index/introducing-stargate-norway/

Jensen Huang congratulates OpenAI on Stargate Norway. It will run on GB300 Superchips, scaling to hundreds of thousands of GPUs – purpose-built for training, reasoning, and real-time inference. “Just as electricity and the internet became foundational to modern life, AI will https://x.com/vitrupo/status/1950828090260955165

🚀We’re expanding the Tencent Hunyuan open-source LLM ecosystem with four compact models (0.5B, 1.8B, 4B, 7B)! Designed for low-power scenarios like consumer-grade GPUs, smart vehicles, smart home devices, mobile phones, and PCs, these models support cost-effective fine-tuning https://x.com/TencentHunyuan/status/1952262079051940322

RT @ori_press: We just benchmarked Qwen 3 Coder and GLM 4.5 on AlgoTune, and they manage to beat Claude Opus 4! We’re excited to see if the…”” / X https://x.com/OfirPress/status/1952470237947085146

ByteDance’s SeedProver scores 331/657 on PutnamBench, almost 4 times the previous SOTA. More impressively, it gets 201/657 under the *light* inference setting, ie equivalent to pass@64-256. DeepSeek-Prover-V2 is just 3 months old… Things go fast now. https://x.com/teortaxesTex/status/1951875052967739787

rule number one: never distill from DeepSeek https://x.com/jxmnop/status/1953163073612562851

wanted to share this Kimi-K2 follow-up because it’s just… so Kimi. Q: It is said that “”great things are best seen at a distance””. Do the Chinese see their country in this world-historical way at all, or do they, like Americans, primarily perceive it from a pedestrian’s point of https://x.com/teortaxesTex/status/1953280294758310310

Announcing Stargate Norway: https://x.com/gdb/status/1950785240122974412

DeepSeek-R1: 2.66 million H800 hours GPT-OSS-120B: 2.1 million H100 hours https://x.com/scaling01/status/1952784655838564376

🚀 Qwen3-30B-A3B-2507 and Qwen3-235B-A22B-2507 now support ultra-long context—up to 1 million tokens! 🔧 Powered by: • Dual Chunk Attention (DCA) – A length extrapolation method that splits long sequences into manageable chunks while preserving global coherence. • https://x.com/Alibaba_Qwen/status/1953760230141309354

Qwen3-Coder is now available on Cerebras, 17x faster than on GPU providers. And it’s completely free. Try it out directly in your developer flow, or signup for our virtual hackathon tomorrow. It’s a $5,000 prize 🙂 @CerebrasSystems @cline https://x.com/SarahChieng/status/1951453803905163693

Small but mighty! Qwen3-Coder-Flash and GLM-4.5-Air are now on @FireworksAI_HQ Despite being smaller and faster, Qwen3 Coder Flash 30B and GLM 4.5-Air achieve almost the same quality as their larger counterparts on tool use benchmarks. The secret of good model behavior is in https://x.com/dzhulgakov/status/1952049826067050735

🚀 Meet Qwen-Image — a 20B MMDiT model for next-gen text-to-image generation. Especially strong at creating stunning graphic posters with native text. Now open-source. 🔍 Key Highlights: 🔹 SOTA text rendering — rivals GPT-4o in English, best-in-class for Chinese 🔹 In-pixel https://x.com/Alibaba_Qwen/status/1952398250121756992

💡 You get 2,000 free Qwen Code runs every day! Run this one simple command: npx @​qwen-code/qwen-code@latest Hit Enter, and that’s it! 🚀 Now with Qwen OAuth support — super easy to use. Try it now and supercharge your vibe code! 💻⚡ Github: https://x.com/Alibaba_Qwen/status/1953835877555151134

Just included example scripts for aligning models using GSPO (including VLM example) 🙆‍♂️🙆‍♂️ GSPO is the latest RL alignment algo by @Alibaba_Qwen and it’s already supported in the latest TRL v0.20 release. Super-easy-to-get-started example scripts below, GO run them! 👩‍💻👩‍💻 https://x.com/SergioPaniego/status/1952305247411691871

Qwen-Image demo on Hugging Face getting absolutely hammered right now 😀 https://x.com/victormustar/status/1952416615351366033

Qwen-Image: Crafting with Native Text Rendering | Qwen https://qwenlm.github.io/blog/qwen-image/

RT @Alibaba_Qwen: 🚀 Introducing Qwen3-4B-Instruct-2507 & Qwen3-4B-Thinking-2507 — smarter, sharper, and 256K-ready! 🔹 Instruct: Boosted ge…”” / X https://x.com/NandoDF/status/1953223478087143640

RT @Alibaba_Qwen: 🚀 Meet Qwen-Image — a 20B MMDiT model for next-gen text-to-image generation. Especially strong at creating stunning graph…”” / X https://x.com/mervenoyann/status/1952455331205841261

So, I did some coding this week… – Qwen3 Coder Flash (30B-A3B) – Mixture-of-Experts setup with 128 experts, 8 active per token – In pure PyTorch (optimized for human readability) – in a standalone Jupyter notebook – Runs on a single A100 https://x.com/rasbt/status/1951635208375034191

Today we release the APIs of our Flash series, which support Qwen3-Coder and Qwen3-2507 now. Both APIs support the context length of 1M tokens. They are fast and accurate, and they are cost-effectve as well. Feel free to take a try! Qwen3-Coder-Flash Model Card:”” / X https://x.com/Alibaba_Qwen/status/1952767585596145773

@ostrisai The VAE is a fine-tune from the Wan 2.1 VAE for image generation, which is super cool and shows how open source foster collaboration, even between rival labs”” / X https://x.com/multimodalart/status/1952409238413684901

China’s cranking out humanoids. NOETIX delivered 105 robots in July alone. https://x.com/TheHumanoidHub/status/1951330695986729435

Trending

Discover more from Ethan B. Holland

Subscribe now to keep reading and get access to the full archive.

Continue reading