Image created with Flux Pro v1.1 Ultra. Image prompt: International, globe assembled from latitude and longitude lines of small bananas, soft shadowed sphere, photorealistic, editorial, minimal, high detail, 3:2 landscape

China’s DeepSeek Preps AI Agent for End-2025 to Rival OpenAI https://finance.yahoo.com/news/china-deepseek-preps-ai-agent-152907224.html?guccounter=1&guce_referrer=aHR0cHM6Ly93d3cuZ29vZ2xlLmNvbS8&guce_referrer_sig=AQAAADiJs67uOGL7PzqX3MGgvD-A6UJVzmztcJfvPzJTz9iF2iWfg-h2zg2pcwJoIuJ-4IUs3BMrEvPbbpbf4j7qXCmM4BqK78UMZVzrZl3fSuokrWneWMYpy8S7L3-xciC9d74km3boS_g57OxikNZN7Owozd204A5KlQA0MSzkqp42

Alibaba shares jump 19% on cloud unit growth, report of new AI chip https://www.cnbc.com/2025/09/01/alibaba-shares-hong-kong-today.html

Alibaba reportedly developing new AI chip as China’s Xi rejects AI’s ‘Cold War mentality’ | Euronews https://www.euronews.com/next/2025/09/01/alibaba-reportedly-developing-new-ai-chip-as-chinas-xi-rejects-ais-cold-war-mentality

USO: Unified Style and Subject-Driven Generation via Disentangled and Reward Learning https://bytedance.github.io/USO/

Unified Style and Subject-Driven Generation via Disentangled and Reward Learning
Great thanks to @_akhaliq for sharing ! 🌟USO is open-sourced and supports you in combining any subjects with any styles in any scenarios! 🚀Give it a try in our demo. 👇👇👇 🥰code https://x.com/fenfenfenfenfan/status/1961464402550690007

“To test models’ performance on Claude Code, we ran GLM-4.5 against Claude Sonnet 4 and other open-source models on 52 practical programming tasks. While GLM-4.5 demonstrated strong performance against top open-source models, it secured a 40.4% win rate against Claude Sonnet 4. https://x.com/Zai_org/status/1962522761630482700

🚀 Introducing slime v0.1.0 — An open-source RL infra powering models like GLM-4.5, built by THUDM & Zhipu AI. @Zai_org RL infra 朱小霖 shared a deep dive on Zhihu into how they redefined high-performance RL infra👇 🛠️ What’s new in v0.1.0? • High-performance inference for https://x.com/ZhihuFrontier/status/1962751555591086226

Announcing GLM Coding Plan for Claude Code! After seeing the amazing adoption of GLM-4.5 over the past month, we’re making it more accessible. Get started: https://x.com/Zai_org/status/1962522757536887205

Have been tinkering with GLM 4.5 for about an hour. It is about 3x faster than Claude Code + Opus 4.1 and 5x faster than GPT-5-high, but still feels just as good as closed-source models. I am definitely more productive than with other models due to GLM-4.5’s speed.”” / X https://x.com/Tim_Dettmers/status/1962603940291260533

Duolingo is facing an existential crisis as Google Translate rolls out features to tutor users—and even handle live translation as a bonus | Fortune https://fortune.com/2025/08/27/duolingo-existential-crisis-ai-google-translate-language-learning-live-translation/

OpenAI Plans to Build Data Center in India in Major Stargate Expansion in Asia – Bloomberg https://www.bloomberg.com/news/articles/2025-09-01/openai-plans-india-data-center-in-major-stargate-expansion?srnd=phx-technology&embedded-checkout=true

Chinese news outlet CCTV Finance: “”According to market data, China’s humanoid robot sales in 2025 will exceed 10,000 units, a year-over-year increase of 125%.”” https://x.com/TheHumanoidHub/status/1961110406858199528

I trained a Qwen Image Edit LoRA for inpainting. Just paint the part you want inpainted green (0, 255, 0), and it will inpaint only that section. https://x.com/ostrisai/status/1963269597865599425

Just forward it to Manus 📩 Inbox overload? Forward it away With Mail Manus, you can offload complicated tasks straight from your inbox. 📩 Forward long newsletters → get summaries 📩 Forward messy threads → get research & next steps 📩 Forward meeting notes → get https://x.com/ManusAI_HQ/status/1961081338582143360

Another major Perplexity iOS app update. Team cooked. Answers are now streamed smooth as butter. Tables, markdown, intermediate steps. Update and enjoy! https://x.com/AravSrinivas/status/1963758210281882029

Pro users in South Korea, Brazil, and Spain can now download Comet. https://x.com/perplexity_ai/status/1963638853975040456

🚀 Select PayPal and @Venmo customers can skip the waitlist for early access to @perplexity_ai’s AI-powered Comet browser and receive a free 12-month Perplexity Pro trial. This offer is part of the new PayPal Subscriptions Hub, where you can: ✨ Manage subscriptions ✨ Update https://x.com/PayPal/status/1963229273071698199

We are rolling out Comet to all students worldwide. Ask Comet to manage your schedule, order textbooks, or prepare for exams with Study Mode. https://x.com/perplexity_ai/status/1963285255198314951

Now Open — AWS Asia Pacific (New Zealand) Region | AWS News Blog https://aws.amazon.com/fr/blogs/aws/now-open-aws-asia-pacific-new-zealand-region/

年初想提升 tool-calling 时特别缺靠谱的benchmark,以为 “”mcp 火了等几天肯定有开源的mcp-bench可用””,结果等了几个月也没等到,但是这最近怎么每周都有好几个 mcp-bench release出来?”” / X https://x.com/bigeagle_xd/status/1961461441799852128

🚨 Top 10 Leaderboard Disrupted ⚡ DeepSeek V3.1 and DeepSeek v3.1 thinking by @deepseek_ai have landed in the Arena, both ranked at #8. A few highlights: 💠 DeepSeek V3.1 is in the Top 3 for Math, Creative Writing & Longer Query 💠 DeepSeek V3.1 thinking comes in #3 for https://x.com/lmarena_ai/status/1961474406817173602

Anyway, here’s a simple fix for the issue. It deviates from the original benchmark, but at least now my silly baseline isn’t better than Qwen3 🤠 For the curious, @akseljoonas and I found this by manually reading the agent trajectories – yet another example where LOOKING AT THE”” / X https://x.com/_lewtun/status/1962884902363255165

🇸🇪 Together AI now has GPU infrastructure located in Sweden – Lower latency across Europe – EU data residency & compliance – GPU clusters + endpoints on demand – Serverless API for GPT-OSS, DeepSeek, Llama, Qwen https://x.com/togethercompute/status/1963498998720872686

✍️ When it comes to creative writing optimization, you can’t ignore Zhi-Create-Qwen3-32B, a fine-tuned variant of Qwen3-32B. On WritingBench, it scores 82.08, outperforming the base model (78.97), showing notable gains across 6 domains (Fig.1) What powers its performance boost? https://x.com/ZhihuFrontier/status/1963441300692402659

Japan Post Bank to launch blockchain-based digital yen https://www.techinasia.com/news/japan-post-bank-to-launch-blockchain-based-digital-yen

Introducing EmbeddingGemma🎉 🔥With only 308M params, this is the top open model under 500M 🌏Trained on 100+ languages 🪆Flexible embeddings (768 to 128 dims) with Matryoshka 🤗Works with your favorite open tools 🤏Runs with as little as 200MB https://x.com/osanseviero/status/1963635281032040914

Tutorial: Train a Qwen Image Edit LoRA with AI Toolkit
https://x.com/ostrisai/status/1961884211956400358

🚀 Kimi K2-0905 just landed on Together AI! Enhanced coding capabilities + 256K context (doubled!). Same open-source power from @Kimi_Moonshot that beats proprietary models. Built for agents. Priced for scale. 👆 https://x.com/togethercompute/status/1963806032548843865

Wonder how the White House feels watching this? > US doubles tariffs on Indian imports (punishing purchases of Russian oil). > Modi responds by shaking hands with Xi, despite a history of border disputes (and all out war) with China. The great game continues…”” / X https://x.com/bilawalsidhu/status/1962242039099388375

Try Hunyuan-MT-7B and Hunyuan-MT-Chimera via @huggingface and @gradio! This model is specialized for translate 🤗”” / X https://x.com/SOSOHAJALAB/status/1962790133054480600

TikTok owner ByteDance sets valuation at over $330 billion in planned buyback – The Japan Times https://www.japantimes.co.jp/business/2025/08/28/tech/tiktok-bytedance-valuation-330-billion/

For llama.vim the recommended setup now is Qwen 3 Coder 30B A3B Instruct: brew install llama.cpp llama-server –fim-qwen-30b-default Amazingly, on Macs the 30B MoE model performs better than the old Qwen 2.5 Coder 7B so if you have the necessary RAM it’s better to switch to https://x.com/ggerganov/status/1961471397428883882

Le Chat. Custom MCP connectors. Memories. | Mistral AI https://mistral.ai/news/le-chat-mcp-connectors-memories

From payments data and refunds to invoices and subscriptions, @MistralAI’s users can now handle it all inside Le Chat with @stripe’s MCP. Here’s how it works: https://x.com/emilygsands/status/1962884010289590583

Tencent open sources two high-performing translation models https://the-decoder.com/tencent-open-sources-two-high-performing-translation-models/

Tencent released Hunyuan-MT-7B and Hunyuan-MT-Chimera The Hunyuan Translation Model comprises a translation model, Hunyuan-MT-7B, and an ensemble model, Hunyuan-MT-Chimera. The translation model is used to translate source text into the target language, while the ensemble model https://x.com/_akhaliq/status/1962644501605835140

Unfortunate reality: most open-source LLM servers (e.g. Together) don’t offer cache-hit discounts, while closed providers like OpenAI do. DeepSeek does discount, but most third-party servers don’t.
https://x.com/arankomatsuzaki/status/1963294646957957263

Hermes 4: Nous Research Open-Weight Reasoning Family Models – 70B / 405B (Llama-3.1 bases, released) – 14B (Qwen3 base, research baseline) Hermes 4 70B & 405B – Base: Llama-3.1-70B / 405B – Training: TorchTitan (modified), Axolotl, 192× B200s, FSDP and TP – Dataset: 56B tokens https://x.com/gm8xx8/status/1962943078702186627

🚀 Qwen-Max has successfully scaled to 1T parameters, and we’re still pushing further. Hopefully this giant will bring some surprises, see you next week!”” / X https://x.com/huybery/status/1963998518667776250

Big news: Introducing Qwen3-Max-Preview (Instruct) — our biggest model yet, with over 1 trillion parameters! 🚀 Now available via Qwen Chat & Alibaba Cloud API. Benchmarks show it beats our previous best, Qwen3-235B-A22B-2507. Internal tests + early user feedback confirm: https://x.com/Alibaba_Qwen/status/1963991502440562976

Qwen3 Max is truly, solidly, a US-grade modern frontier model. They ask $15/MT for what they serve because that is easily its weight class.”” / X https://x.com/teortaxesTex/status/1963994291765649716

Qwen3-Max-Preview is now live on OpenRouter! 🚀”” / X https://x.com/Alibaba_Qwen/status/1964004112149754091

Ready to meet the biggest, brainiest guy in the Qwen3 family?”” / X https://x.com/Alibaba_Qwen/status/1963586344355053865

Really liking the chainlit open source lib for building a quick but nice chat interface for any LLM. Here are some quick single and multi-turn examples for my Qwen3 from-scratch models: https://x.com/rasbt/status/1962695306757185647

Traditional code embedding models face a fundamental bottleneck: there simply aren’t enough high-quality comment-code pairs for supervised training. By starting with Qwen2.5-Coder pre-trained on 5.5 trillion tokens spanning 92+ programming languages, we inherit deep semantic https://x.com/JinaAI_/status/1963637139037720995

Here’s a fun fact about TAU Bench: if you train an SFT baseline which has zero tool-calling capabilities, you can beat Qwen3-4B-Instruct by a large margin on the Airline domain 🙃 Why? Because on this domain, TAU Bench only evaluates the model’s ability to: – communicate with https://x.com/_lewtun/status/1962884893718761634

Glad to see Qwen3-Coder performing well on the GSO leaderboard!”” / X https://x.com/Alibaba_Qwen/status/1963049864474120475

MiniCPM-V 4.5 achieves an average score of 77.0 on OpenCompass, a comprehensive evaluation of 8 popular benchmarks. With only 8B parameters, it surpasses widely used proprietary models like GPT-4o-latest, Gemini-2.0 Pro, and strong open-source models like Qwen2.5-VL 72B powered https://x.com/_akhaliq/status/1963587749400727980

🚨 Attention: Draw Things now officially supports Qwen-Image-Edit.
https://x.com/drawthingsapp/status/1961977481860419771

Huge thanks to the community for making Qwen Image Edit’s inpainting magic happen!🙌
https://x.com/Alibaba_Qwen/status/1963048659676979559

Goated FAIR team just found how coding agents sometimes “”cheat”” on SWE-Bench Verified. It’s really simple. For example, Qwen3 literally greps all commit logs for the issue number of the issue it needs to fix. lol, clever model. “”cheat”” cuz it’s more like env hacking. https://x.com/giffmana/status/1963327672827687316

China’s AgiBot launches sales of six robot types on JD.com https://www.techinasia.com/news/chinas-agibot-launches-sales-robot-types-jdcom

Trending

Discover more from Ethan B. Holland

Subscribe now to keep reading and get access to the full archive.

Continue reading