Image created with Flux Pro v1.1 Ultra. Image prompt: International, globe assembled from latitude and longitude lines of small bananas, soft shadowed sphere, photorealistic, editorial, minimal, high detail, 3:2 landscape
China’s DeepSeek Preps AI Agent for End-2025 to Rival OpenAI https://finance.yahoo.com/news/china-deepseek-preps-ai-agent-152907224.html?guccounter=1&guce_referrer=aHR0cHM6Ly93d3cuZ29vZ2xlLmNvbS8&guce_referrer_sig=AQAAADiJs67uOGL7PzqX3MGgvD-A6UJVzmztcJfvPzJTz9iF2iWfg-h2zg2pcwJoIuJ-4IUs3BMrEvPbbpbf4j7qXCmM4BqK78UMZVzrZl3fSuokrWneWMYpy8S7L3-xciC9d74km3boS_g57OxikNZN7Owozd204A5KlQA0MSzkqp42
Alibaba shares jump 19% on cloud unit growth, report of new AI chip https://www.cnbc.com/2025/09/01/alibaba-shares-hong-kong-today.html
Alibaba reportedly developing new AI chip as China’s Xi rejects AI’s ‘Cold War mentality’ | Euronews https://www.euronews.com/next/2025/09/01/alibaba-reportedly-developing-new-ai-chip-as-chinas-xi-rejects-ais-cold-war-mentality
USO: Unified Style and Subject-Driven Generation via Disentangled and Reward Learning https://bytedance.github.io/USO/
Unified Style and Subject-Driven Generation via Disentangled and Reward Learning
Great thanks to @_akhaliq for sharing ! 🌟USO is open-sourced and supports you in combining any subjects with any styles in any scenarios! 🚀Give it a try in our demo. 👇👇👇 🥰code https://x.com/fenfenfenfenfan/status/1961464402550690007
“To test models’ performance on Claude Code, we ran GLM-4.5 against Claude Sonnet 4 and other open-source models on 52 practical programming tasks. While GLM-4.5 demonstrated strong performance against top open-source models, it secured a 40.4% win rate against Claude Sonnet 4. https://x.com/Zai_org/status/1962522761630482700
🚀 Introducing slime v0.1.0 — An open-source RL infra powering models like GLM-4.5, built by THUDM & Zhipu AI. @Zai_org RL infra 朱小霖 shared a deep dive on Zhihu into how they redefined high-performance RL infra👇 🛠️ What’s new in v0.1.0? • High-performance inference for https://x.com/ZhihuFrontier/status/1962751555591086226
Announcing GLM Coding Plan for Claude Code! After seeing the amazing adoption of GLM-4.5 over the past month, we’re making it more accessible. Get started: https://x.com/Zai_org/status/1962522757536887205
Have been tinkering with GLM 4.5 for about an hour. It is about 3x faster than Claude Code + Opus 4.1 and 5x faster than GPT-5-high, but still feels just as good as closed-source models. I am definitely more productive than with other models due to GLM-4.5’s speed.”” / X https://x.com/Tim_Dettmers/status/1962603940291260533
Duolingo is facing an existential crisis as Google Translate rolls out features to tutor users—and even handle live translation as a bonus | Fortune https://fortune.com/2025/08/27/duolingo-existential-crisis-ai-google-translate-language-learning-live-translation/
OpenAI Plans to Build Data Center in India in Major Stargate Expansion in Asia – Bloomberg https://www.bloomberg.com/news/articles/2025-09-01/openai-plans-india-data-center-in-major-stargate-expansion?srnd=phx-technology&embedded-checkout=true
Chinese news outlet CCTV Finance: “”According to market data, China’s humanoid robot sales in 2025 will exceed 10,000 units, a year-over-year increase of 125%.”” https://x.com/TheHumanoidHub/status/1961110406858199528
I trained a Qwen Image Edit LoRA for inpainting. Just paint the part you want inpainted green (0, 255, 0), and it will inpaint only that section. https://x.com/ostrisai/status/1963269597865599425
Just forward it to Manus 📩 Inbox overload? Forward it away With Mail Manus, you can offload complicated tasks straight from your inbox. 📩 Forward long newsletters → get summaries 📩 Forward messy threads → get research & next steps 📩 Forward meeting notes → get https://x.com/ManusAI_HQ/status/1961081338582143360
Another major Perplexity iOS app update. Team cooked. Answers are now streamed smooth as butter. Tables, markdown, intermediate steps. Update and enjoy! https://x.com/AravSrinivas/status/1963758210281882029
Pro users in South Korea, Brazil, and Spain can now download Comet. https://x.com/perplexity_ai/status/1963638853975040456
🚀 Select PayPal and @Venmo customers can skip the waitlist for early access to @perplexity_ai’s AI-powered Comet browser and receive a free 12-month Perplexity Pro trial. This offer is part of the new PayPal Subscriptions Hub, where you can: ✨ Manage subscriptions ✨ Update https://x.com/PayPal/status/1963229273071698199
We are rolling out Comet to all students worldwide. Ask Comet to manage your schedule, order textbooks, or prepare for exams with Study Mode. https://x.com/perplexity_ai/status/1963285255198314951
Now Open — AWS Asia Pacific (New Zealand) Region | AWS News Blog https://aws.amazon.com/fr/blogs/aws/now-open-aws-asia-pacific-new-zealand-region/
年初想提升 tool-calling 时特别缺靠谱的benchmark,以为 “”mcp 火了等几天肯定有开源的mcp-bench可用””,结果等了几个月也没等到,但是这最近怎么每周都有好几个 mcp-bench release出来?”” / X https://x.com/bigeagle_xd/status/1961461441799852128
🚨 Top 10 Leaderboard Disrupted ⚡ DeepSeek V3.1 and DeepSeek v3.1 thinking by @deepseek_ai have landed in the Arena, both ranked at #8. A few highlights: 💠 DeepSeek V3.1 is in the Top 3 for Math, Creative Writing & Longer Query 💠 DeepSeek V3.1 thinking comes in #3 for https://x.com/lmarena_ai/status/1961474406817173602
Anyway, here’s a simple fix for the issue. It deviates from the original benchmark, but at least now my silly baseline isn’t better than Qwen3 🤠 For the curious, @akseljoonas and I found this by manually reading the agent trajectories – yet another example where LOOKING AT THE”” / X https://x.com/_lewtun/status/1962884902363255165
🇸🇪 Together AI now has GPU infrastructure located in Sweden – Lower latency across Europe – EU data residency & compliance – GPU clusters + endpoints on demand – Serverless API for GPT-OSS, DeepSeek, Llama, Qwen https://x.com/togethercompute/status/1963498998720872686
✍️ When it comes to creative writing optimization, you can’t ignore Zhi-Create-Qwen3-32B, a fine-tuned variant of Qwen3-32B. On WritingBench, it scores 82.08, outperforming the base model (78.97), showing notable gains across 6 domains (Fig.1) What powers its performance boost? https://x.com/ZhihuFrontier/status/1963441300692402659
Japan Post Bank to launch blockchain-based digital yen https://www.techinasia.com/news/japan-post-bank-to-launch-blockchain-based-digital-yen
Introducing EmbeddingGemma🎉 🔥With only 308M params, this is the top open model under 500M 🌏Trained on 100+ languages 🪆Flexible embeddings (768 to 128 dims) with Matryoshka 🤗Works with your favorite open tools 🤏Runs with as little as 200MB https://x.com/osanseviero/status/1963635281032040914
Tutorial: Train a Qwen Image Edit LoRA with AI Toolkit
https://x.com/ostrisai/status/1961884211956400358
🚀 Kimi K2-0905 just landed on Together AI! Enhanced coding capabilities + 256K context (doubled!). Same open-source power from @Kimi_Moonshot that beats proprietary models. Built for agents. Priced for scale. 👆 https://x.com/togethercompute/status/1963806032548843865
Wonder how the White House feels watching this? > US doubles tariffs on Indian imports (punishing purchases of Russian oil). > Modi responds by shaking hands with Xi, despite a history of border disputes (and all out war) with China. The great game continues…”” / X https://x.com/bilawalsidhu/status/1962242039099388375
Try Hunyuan-MT-7B and Hunyuan-MT-Chimera via @huggingface and @gradio! This model is specialized for translate 🤗”” / X https://x.com/SOSOHAJALAB/status/1962790133054480600
TikTok owner ByteDance sets valuation at over $330 billion in planned buyback – The Japan Times https://www.japantimes.co.jp/business/2025/08/28/tech/tiktok-bytedance-valuation-330-billion/
For llama.vim the recommended setup now is Qwen 3 Coder 30B A3B Instruct: brew install llama.cpp llama-server –fim-qwen-30b-default Amazingly, on Macs the 30B MoE model performs better than the old Qwen 2.5 Coder 7B so if you have the necessary RAM it’s better to switch to https://x.com/ggerganov/status/1961471397428883882
Le Chat. Custom MCP connectors. Memories. | Mistral AI https://mistral.ai/news/le-chat-mcp-connectors-memories
From payments data and refunds to invoices and subscriptions, @MistralAI’s users can now handle it all inside Le Chat with @stripe’s MCP. Here’s how it works: https://x.com/emilygsands/status/1962884010289590583
Tencent open sources two high-performing translation models https://the-decoder.com/tencent-open-sources-two-high-performing-translation-models/
Tencent released Hunyuan-MT-7B and Hunyuan-MT-Chimera The Hunyuan Translation Model comprises a translation model, Hunyuan-MT-7B, and an ensemble model, Hunyuan-MT-Chimera. The translation model is used to translate source text into the target language, while the ensemble model https://x.com/_akhaliq/status/1962644501605835140
Unfortunate reality: most open-source LLM servers (e.g. Together) don’t offer cache-hit discounts, while closed providers like OpenAI do. DeepSeek does discount, but most third-party servers don’t.
https://x.com/arankomatsuzaki/status/1963294646957957263
Hermes 4: Nous Research Open-Weight Reasoning Family Models – 70B / 405B (Llama-3.1 bases, released) – 14B (Qwen3 base, research baseline) Hermes 4 70B & 405B – Base: Llama-3.1-70B / 405B – Training: TorchTitan (modified), Axolotl, 192× B200s, FSDP and TP – Dataset: 56B tokens https://x.com/gm8xx8/status/1962943078702186627
🚀 Qwen-Max has successfully scaled to 1T parameters, and we’re still pushing further. Hopefully this giant will bring some surprises, see you next week!”” / X https://x.com/huybery/status/1963998518667776250
Big news: Introducing Qwen3-Max-Preview (Instruct) — our biggest model yet, with over 1 trillion parameters! 🚀 Now available via Qwen Chat & Alibaba Cloud API. Benchmarks show it beats our previous best, Qwen3-235B-A22B-2507. Internal tests + early user feedback confirm: https://x.com/Alibaba_Qwen/status/1963991502440562976
Qwen3 Max is truly, solidly, a US-grade modern frontier model. They ask $15/MT for what they serve because that is easily its weight class.”” / X https://x.com/teortaxesTex/status/1963994291765649716
Qwen3-Max-Preview is now live on OpenRouter! 🚀”” / X https://x.com/Alibaba_Qwen/status/1964004112149754091
Ready to meet the biggest, brainiest guy in the Qwen3 family?”” / X https://x.com/Alibaba_Qwen/status/1963586344355053865
Really liking the chainlit open source lib for building a quick but nice chat interface for any LLM. Here are some quick single and multi-turn examples for my Qwen3 from-scratch models: https://x.com/rasbt/status/1962695306757185647
Traditional code embedding models face a fundamental bottleneck: there simply aren’t enough high-quality comment-code pairs for supervised training. By starting with Qwen2.5-Coder pre-trained on 5.5 trillion tokens spanning 92+ programming languages, we inherit deep semantic https://x.com/JinaAI_/status/1963637139037720995
Here’s a fun fact about TAU Bench: if you train an SFT baseline which has zero tool-calling capabilities, you can beat Qwen3-4B-Instruct by a large margin on the Airline domain 🙃 Why? Because on this domain, TAU Bench only evaluates the model’s ability to: – communicate with https://x.com/_lewtun/status/1962884893718761634
Glad to see Qwen3-Coder performing well on the GSO leaderboard!”” / X https://x.com/Alibaba_Qwen/status/1963049864474120475
MiniCPM-V 4.5 achieves an average score of 77.0 on OpenCompass, a comprehensive evaluation of 8 popular benchmarks. With only 8B parameters, it surpasses widely used proprietary models like GPT-4o-latest, Gemini-2.0 Pro, and strong open-source models like Qwen2.5-VL 72B powered https://x.com/_akhaliq/status/1963587749400727980
🚨 Attention: Draw Things now officially supports Qwen-Image-Edit.
https://x.com/drawthingsapp/status/1961977481860419771
Huge thanks to the community for making Qwen Image Edit’s inpainting magic happen!🙌
https://x.com/Alibaba_Qwen/status/1963048659676979559
Goated FAIR team just found how coding agents sometimes “”cheat”” on SWE-Bench Verified. It’s really simple. For example, Qwen3 literally greps all commit logs for the issue number of the issue it needs to fix. lol, clever model. “”cheat”” cuz it’s more like env hacking. https://x.com/giffmana/status/1963327672827687316
China’s AgiBot launches sales of six robot types on JD.com https://www.techinasia.com/news/chinas-agibot-launches-sales-robot-types-jdcom




