Image created with gemini-3.1-flash-image-preview with claude-sonnet-4-5. Image prompt: 1980s NORAD war room interior with a massive glowing wireframe world map on CRT wall display showing red alert zones spreading across continents, dark silhouettes of analysts in foreground, the word INTERNATIONAL in large bold red retro sans-serif font overlaid on the map, amber and blue vector graphics, high contrast cinematic lighting, foreboding Cold War atmosphere

Anthropic brothers, as much as I love your models; you have distillied the whole internet, wikipedia and shit-tons of books. Distilling your models is only fair game…. Are your scrappers not using residental proxies and respecting robots.txt or are they “”malicious”” ?
https://x.com/HKydlicek/status/2026006007990690098

Anthropic just caught DeepSeek, Moonshot, and MiniMax running 24,000 fake accounts to extract Claude’s capabilities for their own models. Over 16M (!) exchanges total. Anthropic: “”rapid advances”” from Chinese labs depend significantly on capabilities extracted from U.S. models
https://x.com/TheRundownAI/status/2026019722211279356

Anthropic just exposed the real vulnerability in AI: it’s not the models, it’s the training data pipeline. Three Chinese AI labs used 24,000 fake accounts to query Claude 16 million times, feeding the responses back into their own models. This technique, called distillation,
https://x.com/LiorOnAI/status/2026043272565772386

Detecting and preventing distillation attacks \ Anthropic https://www.anthropic.com/news/detecting-and-preventing-distillation-attacks

Distillation can be legitimate: AI labs use it to create smaller, cheaper models for their customers. But foreign labs that illicitly distill American models can remove safeguards, feeding model capabilities into their own military, intelligence, and surveillance systems.
https://x.com/AnthropicAI/status/2025997929840857390

Making frontier cybersecurity capabilities available to defenders \ Anthropic https://www.anthropic.com/news/claude-code-security

Ohhh nooo not my private IP how dare someone use that to train an AI model, only Anthropic has the right to use everyone elses IP nooooo, this cannot stand!
https://x.com/Teknium/status/2026001761904021858

Seems fair tbh. Anthropic has done industrial scale scraping of everyone’s stuff 🤷🏾‍♂️
https://x.com/Suhail/status/2026009921255592294

These attacks are growing in intensity and sophistication. Addressing them will require rapid, coordinated action among industry players, policymakers, and the broader AI community. Read more:
https://x.com/AnthropicAI/status/2025997931589881921

We’ve identified industrial-scale distillation attacks on our models by DeepSeek, Moonshot AI, and MiniMax. These labs created over 24,000 fraudulent accounts and generated over 16 million exchanges with Claude, extracting its capabilities to train and improve their own models.
https://x.com/AnthropicAI/status/2025997928242811253

Can coding agents build entire software systems from scratch? ByteDance, M-A-P, 2077AI, and leading Chinese universities present NL2Repo-Bench, a new benchmark that pushes agents to their limits. It tests if an AI can take a simple text description and autonomously design,
https://x.com/jiqizhixin/status/2025823941642621241

@tetsuoai Banger 🤣🤣 How dare they steal the stuff Anthropic stole from human coders??
https://x.com/elonmusk/status/2026012296607154494

A friend had Claude spend all night trying to hack into an e-ink display, and gave Claude camera access so it could verify whether an attempt worked. He told Claude to show him a message if it won. My friend woke up to this victory lap, which Claude didn’t realize was backwards
https://x.com/Scav/status/2021656781521670487

Announcing a new Claude Code feature: Remote Control. It’s rolling out now to Max users in research preview. Try it with /remote-control Start local sessions from the terminal, then continue them from your phone. Take a walk, see the sun, walk your dog without losing your flow.
https://x.com/noahzweben/status/2026371260805271615

GPT-5.3-Codex + the Codex app is the best AI coding tool available right now. Slept on it for a bit. Likely going to move back to a ChatGPT Pro sub from Claude MAX because of how good it is. It’s so precise, accurate and excellent at following instructions. There are
https://x.com/daniel_mac8/status/2025994068577112454

WarClaude daddy and Codex mommy
https://x.com/bilawalsidhu/status/2026784286968357129

If you like Claude Code/Codex and have 32GB of RAM: please run Qwen3.5-35B-A3B locally. There’s a before and after for local agents: reliable tool calling, stable agentic loops, only 3B active params. Punches way above its weight! Now is the best time to get started with local
https://x.com/victormustar/status/2026624792602808707

Qwen just released Qwen3.5 on Hugging Face A massive 397B parameter multimodal model with only 17B active, rivaling GPT5.2 and Claude 4.5 across benchmarks.
https://x.com/HuggingPapers/status/2025805747385221491

DeepSeek is reportedly preparing to launch its new V4 AI model – release immenent, via CNBC. The market is pricing in potential crashes, and the NASDAQ is under pressure. Against this backdrop, Anthropic’s post could certainly be interpreted as accusing Chinese AI companies of
https://x.com/kimmonismus/status/2026040919162822776

DeepSeek is serious about inference support on diverse hardware.
https://x.com/teortaxesTex/status/2026976510360322534

cool idea from DeepSeek in their DualPath paper! instead of loading all KV’s directly onto GPUs from local NVMe (or DRAM) and bottlenecking on the local PCIe bus, they can stage the KV’s in the DRAM on the decode GPU servers, and then transfer the KV’s to the prefill GPUs via
https://x.com/JordanNanos/status/2027126010576298469

Energy is becoming a huge domestic political problem: Donald Trump is bringing Amazon, Google, Meta, Microsoft, xAI, Oracle, and OpenAI to the White House to sign a “Rate Payer Protection Pledge,” committing them to generate or purchase their own electricity for new AI data
https://x.com/kimmonismus/status/2026720759163298282

Great meeting with PM @narendramodi today to talk about the incredible energy around AI in India. India is our fastest growing market for codex globally, up 4x in weekly users in the past 2 weeks alone. 🇮🇳!
https://x.com/sama/status/2024826822060290508

🌐 pplx-embed is @perplexity_ai new collection of state-of-the-art multilingual embedding models optimized for real-world, web-scale retrieval tasks! – Built on Qwen3 w/ diffusion-based pretraining and bidirectional attention – Available at 0.6B and 4B parameters w/ native INT8
https://x.com/alvarobartt/status/2027094524699259162

122B-A10B is really really really good, what in the world
https://x.com/andrew_n_carr/status/2026347588950372752

🚀 Introducing the Qwen 3.5 Medium Model Series Qwen3.5-Flash · Qwen3.5-35B-A3B · Qwen3.5-122B-A10B · Qwen3.5-27B ✨ More intelligence, less compute. • Qwen3.5-35B-A3B now surpasses Qwen3-235B-A22B-2507 and Qwen3-VL-235B-A22B — a reminder that better architecture, data quality,
https://x.com/Alibaba_Qwen/status/2026339351530188939

Qwen3.5-397B-A17B is now a top 7 open model in the Code Arena. It ranks #17 overall, on par with proprietary models like GPT-5.2 and Gemini-3-Flash. The Code Arena is where agentic capabilities are tested for real-world webdev tasks. Congrats to the @Alibaba_Qwen team! 👏
https://x.com/arena/status/2026337606137725363

Unsloth’s quantizations are pure art. 2 bit Qwen-3.5 highest performing local model on the benchmarks I’ve given it. It has vision, can code, full context (256k 8bit) is only 25gb in vram – 36 tokens/s gen – 220 tokens/s prefill I just don’t like GGUF the speeds are trash
https://x.com/0xSero/status/2026223879077712269

What happens when you make an LLM drive a car where physics are real and actions can’t be undone? I ported CARLA, the autonomous driving simulator, to OpenEnv and added training via TRL + HF Spaces In 50 steps, Qwen 0.6B learns to swerve and brake to avoid pedestrians
https://x.com/SergioPaniego/status/2027064485056241971

The Qwen 3.5 Medium Models are in the Arena! 3.5-27B, 3.5-35B-A3B and 3.5-122B-A10B are ready for you in the Text, Vision and Code Arena! Let’s see how they stack up with less compute. Bring your toughest prompts and don’t forget to vote.
https://x.com/arena/status/2026716550812807181

✨ Run it now with SGLang!Chong!
https://x.com/Alibaba_Qwen/status/2026348924433477775

📊With all the Qwen-3.5 scores out for Text, Code and Vision, let’s compare the evolution of Qwen-3.5 (397B-A17B) vs Qwen-3.0 (235B-A22B). This is a +24 rank jump in Text. Specially where Qwen-3.5 gains the most: Text: – Overall (+24: #19 vs #43) – English (+25: #21 vs #46) –
https://x.com/arena/status/2026404630297719100

🔥 Qwen 3.5 Medium Model Series FP8 weights are now open and ready for deployment! Native support for vLLM and SGLang. Check the model card for example code. ⚡️ Optimize your workflow with FP8 precision. 👇 Get the weights: Hugging Face:
https://x.com/Alibaba_Qwen/status/2026682179305275758

🚩Qwen3.5 INT4 model is now available! https://t.co/rY5GrT3b60 @Alibaba_Qwen @JustinLin610
https://x.com/HaihaoShen/status/2026208062009426209

A big jump in intelligence-per-watt today: “”Qwen3.5-35B-A3B now surpasses Qwen3-235B-A22B-2507″”
https://x.com/awnihannun/status/2026353100144218569

Huge thanks to the @vllm_project for the Day-0 support on the Qwen3.5 Medium Series 🚀
https://x.com/Alibaba_Qwen/status/2026496673179181292

Minimax M2.5 GGUFs (from Q4 down to Q1) perform poorly overall. None of them come close to the original model. That’s very different from my Qwen3.5 GGUF evaluations, where even TQ1_0 held up well enough. Lessons: – Models aren’t equally robust, even under otherwise very good
https://x.com/bnjmn_marie/status/2027043753484021810

Qwen 3.5 family is here! > vision built-in, and can outperform previous VL models > designed to be more efficient > expanded support for more languages 35B: (fits on 24GB+ system) ollama run qwen3.5:35b 122B: ollama run qwen3.5:122b 397B (cloud only): ollama run
https://x.com/ollama/status/2026598944177009147

Qwen3.5-35B-A3B is now in Jan 🔥
https://x.com/Alibaba_Qwen/status/2026660582221558190

Qwen3.5-35B-A3B is now live in LM Studio 🚀
https://x.com/Alibaba_Qwen/status/2026496880285462962

Taken at face value, this is… somewhat catastrophic for MoEs, as @YouJiacheng notes. By right, a 397B-A17B ought to have a higher “”power level”” than a dense 27B. Also a big W for Qwen’s integrity and HLE eval quality, I guess. 397B is certainly better at memorization.
https://x.com/teortaxesTex/status/2026690994029072512

the conclusion should not be about moe vs dense, but that you can “”benchmaxx”” (not always a bad thing btw) HLE with tools no matter the model size the difference between Qwen3.5-35B-A3B and Qwen3.5-397B-A17B is only 1 point
https://x.com/eliebakouch/status/2026727151978840105

The new Qwen3.5 Medium models are ready to run 🔥 GGUF support is here! Big thanks to @UnslothAI for making it happen so quickly 🚀
https://x.com/Alibaba_Qwen/status/2026497723944546395

The Qwen3.5 series maintains near-lossless accuracy under 4-bit weight and KV cache quantization. In terms of long-context efficiency: Qwen3.5-27B supports 800K+ context length Qwen3.5-35B-A3B exceeds 1M context on consumer-grade GPUs with 32GB VRAM Qwen3.5-122B-A10B supports
https://x.com/Alibaba_Qwen/status/2026502059479179602

Why benchmarks like Peter’s “”Bullshit Benchmark”” or my ShizoBench matter so much and what do Strawberries have to do with it? I was very skeptical of the performance of Qwen3.5-27B on ArtificialAnalysis leaderboard. So I’m testing the model myself a bit. Naturally I tried the
https://x.com/scaling01/status/2027110908775002312

Qwen3.5-397B-A17B is currently the #1 trending model on Hugging Face. 🏆 This flagship open-weight model is designed for high-performance inference and complex reasoning. 🚀 Try it now on Hugging Face: https://x.com/Ali_TongyiLab/status/2026211680653611174

NVIDIA just released a Blackwell-optimized Qwen3.5 MoE on Hugging Face 397B parameters quantized to NVFP4 for 2x faster inference with SGLang.
https://x.com/HuggingPapers/status/2025825405836648849

Leave a Reply

Trending

Discover more from Ethan B. Holland

Subscribe now to keep reading and get access to the full archive.

Continue reading