Image created with gemini-3.1-flash-image-preview with claude-sonnet-4-5. Image prompt: Using the provided reference image hoodornament.jpg, preserve the deep midnight navy car hood, shallow depth-of-field sky background, chrome pedestal base, dramatic upward camera angle, and automotive advertisement lighting exactly as shown. Replace only the Mercedes star with a single chrome open padlock hood ornament, shackle swung open to one side, rendered in the same polished metal at realistic ornament scale, mounted on the identical pedestal. Add bold white sans-serif display text reading OPEN SOURCE across the upper portion of the image as a clean headline.

Introducing Forge | Mistral AI https://mistral.ai/news/forge

Introducing Mistral Small 4 | Mistral AI https://mistral.ai/news/mistral-small-4

Leanstral: Open-Source foundation for trustworthy vibe-coding | Mistral AI https://mistral.ai/news/leanstral

@_avichawla Impressive work from Kimi
https://x.com/elonmusk/status/2033528245464047805

🔥 @Kimi_Moonshot’s new Attention Residual paper is sparking discussions. Zhihu contributor OpenLLMAI shares a deep dive: “”From Kimi’s Attention Residual to ‘Vertical Attention’ — an idea I’ve been thinking about for half a year.”” Some interesting thoughts on attention mechanisms
https://x.com/ZhihuFrontier/status/2033751367198949865

Avi Chawla on X: “Big release from Kimi! They just released a new way to handle residual connections in Transformers. In a standard Transformer, every sub-layer (attention or MLP) computes an output and adds it back to the input via a residual connection. If you consider this across 40+ layers, https://t.co/5i5AN9tzIm” / X
https://x.com/_avichawla/status/2033472650836914495

https://chatgpt.com/share/69cda240-9324-832a-89b6-a43d4a22f437

https://claude.ai/share/7239e73e-9e9d-469a-bbdb-e5c7da75a4e9

Introducing 𝑨𝒕𝒕𝒆𝒏𝒕𝒊𝒐𝒏 𝑹𝒆𝒔𝒊𝒅𝒖𝒂𝒍𝒔: Rethinking depth-wise aggregation. Residual connections have long relied on fixed, uniform accumulation. Inspired by the duality of time and depth, we introduce Attention Residuals, replacing standard depth-wise recurrence with
https://x.com/Kimi_Moonshot/status/2033378587878072424

visual summary of attention residuals by kimi, beautiful paper
https://x.com/eliebakouch/status/2033488233854620007

Announced in Jensen’s keynote today: LangChain frameworks have crossed 1B downloads. We’re excited to join the NVIDIA Nemotron Coalition to help shape the open models that power these agents. ➡️ Read the announcement: https://t.co/CWlbAzhlXy ➡️ Check out the docs:
https://x.com/LangChain/status/2033788913937195132

🙌 Andrej Karpathy’s lab has received the first DGX Station GB300 — a Dell Pro Max with GB300. 💚 We can’t wait to see what you’ll create @karpathy! 🔗 https://t.co/8ct5QZ3frS @DellTech
https://x.com/NVIDIAAIDev/status/2034291235041554871

Jensen is cementing the idea that Nvidia-powered AI is now the backbone of every major industry. He said robotics alone will be a $50 trillion industry.
https://x.com/TheHumanoidHub/status/2033619022508659118

Jensen: “Nvidia is the first vertically integrated but horizontally open company.” This strategy positions Nvidia as the backbone of robotics without stifling innovation. Vertical integration ensures cutting-edge performance on each layer of the AI stack. Horizontal openness
https://x.com/TheHumanoidHub/status/2033622691408974133

The First Healthcare Robotics Dataset and Foundational Physical AI Models for Healthcare Robotics https://huggingface.co/blog/nvidia/physical-ai-for-healthcare-robotics

Announcing NVIDIA DLSS 5, an AI-powered breakthrough in visual fidelity for games, coming this fall. DLSS 5 infuses pixels with photorealistic lighting and materials, bridging the gap between rendering and reality. Learn More → https://x.com/NVIDIAGeForce/status/2033617732147810782

DLSS 5 is completely mind blowing. The neural rendering model with photoreal lighting and materials is a generation step up in visual fidelity. Gaming with DLSS 5 feels like future tech, but its possible now. It is truly incredible. 🤯
https://x.com/GeForce_JacobF/status/2033615891045454112

DLSS 5 might be the moment where the anti AI pendulum starts swinging back. Many in the 3D community who were against generative AI are now pushing back on the “”everything is AI slop”” crowd. The pendulum swung too far and they can feel it. Nice to see the rebalancing.
https://x.com/bilawalsidhu/status/2034281398052274666

Here’s everything we know about Nvidia’s “”greatest leap in graphics since real-time ray tracing”” You can see Digital Foundry’s jaw drop in this reaction after they just saw DLSS 5.0: – Will ship in Fall of 2026! – Demo ran 4k on 2 5090’s but is already running on single GPU in
https://x.com/Grummz/status/2033641075806769382

GR00T is moving away from VLM-based backbones in favor of integrated world models. Jensen Huang teased GR00T N2 during his keynote; NVIDIA’s next-gen foundation model built on DreamZero research. Utilizing a new world-action model architecture, it succeeds at novel tasks in
https://x.com/TheHumanoidHub/status/2034279221372321940

What if a robot could simulate the physical world from a single image. [📍Bookmark Paper & GitHub for later] PointWorld-1B from Stanford and NVIDIA is a large 3D world model that predicts how an entire scene will move, given RGB-D input and robot actions. The key idea is
https://x.com/IlirAliu_/status/2032895393407660380

Breaking: 1 trillion revenue for NVIDIA in 2027 Jensen Huang: “One year after last GTC, right here where I stand… I see, going down so much, through 2027. At least… one trillion dollars, you know? Now, does it make any sense? I’m certain computer demand will be much
https://x.com/TheTuringPost/status/2033622628385362068

Jensen just said NVIDIA’s $1T projection for 2025-27 covers only Blackwell and Rubin to keep it consistent with the previous projection. He mentioned he could have included Groq in that number: “”so if I would’ve included that, theoretically, not actually, but theoretically,
https://x.com/TheHumanoidHub/status/2033990614824665421

Nvidia targets data center revenue of $1+ trillion for 2025-2027. That’s already quite ridiculous, with the AI physical world only in its zeroth innings . $NVDA
https://x.com/TheHumanoidHub/status/2033627322331660784

A breakthrough in real-time video generation. As a research preview developed with @NVIDIA and shared at @NVIDIAGTC this week, we trained a new real-time video model running on Vera Rubin. HD videos generate instantly, with time-to-first-frame under 100ms. Unlocking an entirely
https://x.com/runwayml/status/2034284298769985914#m

NVIDIA GTC 2026 Keynote: Everything That Happened in 12 Minutes – YouTube https://www.youtube.com/watch?v=X2i_8O75_Os

Introducing LangSmith Fleet: an enterprise workspace for creating, using, and managing your fleet of agents. Fleet agents have their own memory, access to a collection of tools and skills, and can be exposed through the communication channels your team uses every day. Fleet
https://x.com/LangChain/status/2034679590250258855

Introducing LangSmith Fleet. Agents for every team. → Build agents with natural language → Share and control who can edit, run, or clone each agent → Manage authentication with agent identity → Approve actions with human-in-the-loop → Track and audit actions with tracing in
https://x.com/LangChain/status/2034694530478612777

LangChain just open-sourced a replica of Claude Code. It’s called Deep Agents. MIT licensed, model-agnostic, and fully inspectable – so you can finally see exactly how coding agents like Claude Code are built under the hood. The black box just became a textbook. GitHub:
https://x.com/RoundtableSpace/status/2033955271333011829

LangChain just open-sourced Deep Agents–an agent harness that’s opinionated and ready-to-run out of the box. Instead of wiring up prompts, tools, and context management yourself, you get a working agent immediately and customize what you need. It’s an MIT-licensed system that’s
https://x.com/itsafiz/status/2033591253955449289

Schibsted open sources AI tool that turns news articles into videos | Schibsted https://schibsted.com/news/schibsted-open-sources-ai-tool-that-turns-news-articles-into-videos/

i am actually still not over how Qwen as we knew it, one of the S tier Tigers, is over.
https://x.com/swyx/status/2033030744352993296

build agents with LangSmith and let them execute code securely. We’re launching LangSmith Sandboxes today!
https://x.com/samecrowder/status/2034123616720421210#m

more and more agents will write and execute code launching langsmith sandboxes (waitlist to start) to make this easy we’ll be letting people off pretty quickly, so sign up! or dm me
https://x.com/hwchase17/status/2033950657619874217

We just made it dramatically easier for agents to read trending research papers on HF. Let’s go AI powered research!
https://x.com/ClementDelangue/status/2034277529981178007#m

We just released an hf CLI extension to detect the best model/quant for a user’s hardware and then spins up a local coding agent. Time to go local/private/free/fast for your agents thanks to open-source!
https://x.com/ClementDelangue/status/2033982183791108278

We’re launching LangSmith Fleet today! There are some primitives in Fleet that I think will be very useful in a future where agents do a lot of the world’s work – Agent Identity: as more work is specified by humans but done by agents, we need identity + security models that
https://x.com/Vtrivedy10/status/2034690067839521114

Leanstral is part of the Mistral Small 4 family
https://x.com/scaling01/status/2033625927268126969

Analysis of training dynamics demonstrates how AttnRes naturally mitigates hidden-state magnitude growth and yields a more uniform gradient distribution across depth.
https://x.com/Kimi_Moonshot/status/2033378596438556853

I wrote something on Moonshot’s latest research release – Attention Residuals. Intuition, notes and how you can understand standard residuals vs mHC vs attention residuals.
https://x.com/tokenbender/status/2033437211371454915

Moonshot AI targets $1b raise, eyes $18b valuation https://www.techinasia.com/news/moonshot-ai-targets-1b-raise-eyes-18b-valuation

This is so damn cool! Transformers do attention across tokens, now imagine doing attention across layers too. This delivers a 1.25x compute efficiency, <4% training overhead on the 48B Kimi model, +7.5 on GPQA-Diamond. Kimi is quietly becoming the new DeepSeek for the coolest
https://x.com/Yuchenj_UW/status/2033404695880896804

Oh wow, Mamba-3 is here! For me, the most interesting use case of Mamba and Mamba-likes are the recent transformer attention hybrid architectures (Qwen3.5, Kimi Linear, etc.) Would be interesting to swap Gated DeltaNet with Mamba-3 (which now also has RoPE) in next gen hybrids.
https://x.com/rasbt/status/2034088726997893168#m

📎We’ve uploaded it to arXiv, enjoy! https://x.com/Kimi_Moonshot/status/2033796781327454686

🔥 An insider take on @Kimi_Moonshot ‘s Attention Residual — From Kimi AI infra team member & Zhihu contributor Reku A rare look at how attention ideas collide with real-world training systems 👇 🧠 Attention Residual isn’t just modeling — it’s an infra challenge I mainly worked
https://x.com/ZhihuFrontier/status/2034269774281400798#m

As a member of the Kimi team, I wrote the linked blog to share how our team tackles truly innovative work together–not just as individuals, but as a coordinated group. 💎I fully agree: “you can always trust the Kimi solidness.” For us, solidness means making ideas actually work
https://x.com/YyWangCS17122/status/2034273847164473820#m

For more details, check out our paper here:
https://x.com/Kimi_Moonshot/status/2033378599450079581

Thread by @Kimi_Moonshot on Thread Reader App – Thread Reader App https://threadreaderapp.com/thread/2033378587878072424.html

Xiaomi has released MiMo-V2-Pro, which scores 49 on the Artificial Analysis Intelligence Index, placing it between Kimi K2.5 and GLM-5 @Xiaomi’s MiMo-V2-Pro is a new reasoning model and a significant upgrade over their prior open weights release, MiMo-V2-Flash (309B total / 15B
https://x.com/ArtificialAnlys/status/2034239267052896516#m

Finishing a video episode of Attention Span about super interesting announcement from #NVIDIAGTC
https://x.com/TheTuringPost/status/2033568823396430101

New Scaling Law? What “Agentic Scaling”” Is – Inside NVIDIA’s Biggest Idea at GTC 2026
https://x.com/TheTuringPost/status/2033689291419734102

NVIDIA’s Nemotron 3 is an architectural response to the 2 pressures: – Long-context cost as agentic interactions scale – Repeated reasoning cost from invoking full models for small subtasks Nemotron 3 proposes several design decisions to solve this: ▪️ Hybrid architecture:
https://x.com/TheTuringPost/status/2034668980892479993

NemoClaw – NVIDIA’s contribution to the emerging OpenClaw ecosystem and one of the biggest announcements at NVIDIA GTC It’s a framework for long-running autonomous agents. ▪️ The idea: Install OpenClaw together with Nemotron models and OpenShell (NVIDIA’s new security runtime)
https://x.com/TheTuringPost/status/2034389444875428043

💚🤗💚 Jensen showing @huggingface during GTC keynote, where @NVIDIAAI dropped amazing new open models, datasets and blogs! Some of my favorites, links in comments: 🧠 Nemotron 3 Super 120A12B – Reasoning LLM 🏥 Open-H-Embodiment – Healthcare Robotics Dataset 🩻
https://x.com/jeffboudier/status/2033959279510884631

Jensen Huang: “It is now one of the recruiting tools in Silicon Valley. How many tokens comes along with my job?” @NVIDIAGTC
https://x.com/TheTuringPost/status/2033639746128515518

NVIDIA’s strategy in one picture @NVIDIAGTC
https://x.com/TheTuringPost/status/2033620574694752678

Robotics research is accelerating fast, especially around simulation. Factory deployment still isn’t. The gap between simulation and real production lines remains one of the biggest bottlenecks in manufacturing automation. That’s why @ABBRobotics’s partnership with @NVIDIA
https://x.com/IlirAliu_/status/2033381389232689529

Second day! “Technology Behind Robotic Characters”, session at @nvidia GTC. Moritz Baecher on how @Disney Imagineering builds believable physical AI: Many robotics teams struggle to move from digital animation to stable physical movement. Their approach bridges that gap. The
https://x.com/IlirAliu_/status/2033980181413827053

With legendary @Scobleizer and @wschenk #nvidiagtc @NVIDIAGTC
https://x.com/TheTuringPost/status/2033574233360699881

And 2.3 years later we have DLSS on steroids
https://x.com/bilawalsidhu/status/2033752195095535801

DLSS 5 casually solved the fancy coat of paint part of this vision
https://x.com/bilawalsidhu/status/2034131183353643289

DLSS 6 mode on about to take greyboxed 3d assets to final render. Ai video-to-video foreshadowed this; many said it could never happen in real time. Yet here we are.
https://x.com/bilawalsidhu/status/2033898489952841763

So proud of DLSS5: Fully generative neural rendering, in real-time, in real games. Mind-blowing realism. A whole new generation of real-time graphics. A decade of continuous research and development. Coming soon to PCs everywhere. 💚
https://x.com/ctnzr/status/2033613807105544666

Jensen Huang’s view on autonomous vehicles is pretty straightforward: the “automotive is less than 1% of your business” number misses what is actually happening. NVIDIA is selling three computers: – training systems – simulation and synthetic data – the AV system in the car
https://x.com/TheTuringPost/status/2033992848203514225

Been so much fun cooking OpenShell and NemoClaw with the @NVIDIAAI folks! 🙏🦞 Huge step towards secure agents you can trust. What’s your OpenClaw strategy?
https://x.com/steipete/status/2033641463104323868

GTC 2026 News | NVIDIA Newsroom https://nvidianews.nvidia.com/online-press-kit/gtc-2026-news

Jensen says he can’t think of a company building robots that isn’t working with Nvidia.
https://x.com/TheHumanoidHub/status/2033642974492659894

NVIDIA GTC 2026: Live Updates on What’s Next in AI | NVIDIA Blog https://blogs.nvidia.com/blog/gtc-2026-news/

Developers used to argue about programming languages; now they argue about harnesses. NemoClaw is NVIDIA’s answer to your OpenClaw safety woes — zero permissions by default, sandboxed subagents, private inference enforced at the infra layer. Here’s a guide on how to start:
https://x.com/baseten/status/2034649896523874356

Go from “”hello world”” to “”hello claw!”” 🦞 We’re hosting a Build-A-Claw extravaganza in the #NVIDIAGTC Park Mon-Thur where you can BYOD or buy a DGX Spark on-site and our NVIDIA experts will help you install @OpenClaw. See you there! 🙌 Full details 👉 https://x.com/NVIDIAAIDev/status/2032847578404888907

We’re going live at #NVIDIAGTC in 30 minutes. ⏱️ Join us for GTC Live at 8 a.m. PT as we get ready for Jensen Huang’s keynote 11 a.m. Featuring industry leaders from: @bfl_ml, @Cadence, @CaterpillarInc, @cohere, @CoreWeave, @DellTech, @EdisonSci, @FireworksAI_HQ, @IBM,
https://x.com/nvidia/status/2033551362210865371

🚀 Live from @NVIDIAGTC, we’re releasing Holotron-12B! Developed with @nvidia, it’s a high-throughput, open-source, multimodal model engineered specifically for the age of computer-use agents. Get started today! 🤗Hugging Face: https://t.co/SyAuqLIacS 📖Technical Deep Dive:
https://x.com/hcompany_ai/status/2033851052714320083

AI is already redesigning chip design itself! And the biggest bottleneck left is validation. Here is Bill Dally describing to @JeffDean how @nvidia uses AI to design chips: “We’re already using AI across multiple parts of the chip design process, and it’s delivering real
https://x.com/TheTuringPost/status/2034413469542588613

How NVIDIA Dynamo 1.0 Powers Multi-Node Inference at Production Scale | NVIDIA Technical Blog https://developer.nvidia.com/blog/nvidia-dynamo-1-production-ready/

With Nemotron 3 Nano 4B in the NVIDIA Nemotron 3 family, llama.cpp users get a compact model for action-taking conversational personas, available across NVIDIA GPU-enabled systems and @NVIDIA_AI_PC
https://x.com/ggerganov/status/2033947673825337477

The frontier has increasingly shifted to hybrid models – from Qwen to Kimi-Linear and now with NVIDIA’s Nemotron-3 Super – that rely on a strong linear sequence model. Today we release Mamba-3, the most powerful linear model to date.
https://x.com/tri_dao/status/2033948569502413245

NVIDIA thanks all its partners: the message? There is no way around NVIDIA. NVIDIA is the center of the revolution.
https://x.com/kimmonismus/status/2033615181415387610

Straight from NVIDIA GTC: Jensen Huang just unveiled a new vision for AI infrastructure For the first time, Rubin GPUs+Groq LPUs are paired: > 35× higher inference throughput > 10× more revenue from trillion-parameter models Architecture & why it’s needed
https://x.com/TheTuringPost/status/2033700480975520097

Thank you Jensen and NVIDIA! She’s a real beauty! I was told I’d be getting a secret gift, with a hint that it requires 20 amps. (So I knew it had to be good). She’ll make for a beautiful, spacious home for my Dobby the House Elf claw, among lots of other tinkering, thank you!!
https://x.com/karpathy/status/2034321875506196585

Introducing Unsloth Studio | Unsloth Documentation https://unsloth.ai/docs/new/studio

Introducing Unsloth Studio ✨ A new open-source web UI to train and run LLMs. • Run models locally on Mac, Windows, Linux • Train 500+ models 2x faster with 70% less VRAM • Supports GGUF, vision, audio, embedding models • Auto-create datasets from PDF, CSV, DOCX •
https://x.com/UnslothAI/status/2033926272481718523

Rethinking open source mentorship in the AI era – The GitHub Blog https://github.blog/open-source/maintainers/rethinking-open-source-mentorship-in-the-ai-era/

🚀 Today we’re launching LangSmith Sandboxes Agents get a lot more useful when they can run code: analyze data, call APIs, build entire applications. Sandboxes give them a safe place to do it with ephemeral, locked-down environments you control. Now in Private Preview. Learn
https://x.com/LangChain/status/2033949251529793978

Deploy LangGraph agents using the LangGraph CLI You can now deploy LangGraph agents to production straight from your terminal using the LangGraph CLI! 🛠️ langgraph new → scaffold from a template 🧪 langgraph dev → test locally in Studio 🚀 langgraph deploy → deploy your
https://x.com/LangChain/status/2033596690171629582

Fantastic write-up on SKILLS by the GOAT @trq212 I have been leaning into SKILLS a ton too, both for my Claude Code setup, and also for building agentic software (mostly with DeepAgents from LangChain). You should read his post top to bottom, and if you don’t have much time,
https://x.com/mstockton/status/2034095691648098606#m

New Conceptual Guide: You don’t know what your agent will do until it’s in production 👀 With traditional software, you ship with reasonable confidence. Test coverage handles most paths. Monitoring catches errors, latency, and query issues. When something breaks, you read the
https://x.com/LangChain/status/2034314483259031965#m

Polly is our AI assistant built directly into LangSmith to help you debug, analyze, and improve your agents — now generally available. Now, Polly lives on every page of LangSmith, remembers your full session as you navigate, and can take action to update prompts, compare
https://x.com/LangChain/status/2034321435418825023#m

You can now build your own version of Claude Code. Deep Agents is a MIT-licensed framework that recreates the core workflow behind top coding agents. It lets you inspect and modify the exact architecture that makes these agents work. – Planning and todo tools for managing
https://x.com/simplifyinAI/status/2033581939756818648

I reverse engineered Qwen 3.5’s FP8 format, and provide a script to recreate it.
https://x.com/QuixiAI/status/2033419073401287156

Pretty proud of this one! 😎 Qwen 3.5 Max Preview just hit #3 in Math, Top 10 in Arena Expert, and Top 15 overall! We’re already back in the lab optimizing the preview experience. Even sharper performance coming soon–stay tuned! 🚀
https://x.com/Alibaba_Qwen/status/2034658901321560549

Qwen 3.5 Max Preview has landed in top 10 for Arena Expert and top 15 for Text Arena. It shows particular strength in Math. Highlights: – #3 Math – #10 Expert – #15 Text Arena – Top 20 for Writing, Literature & Language, Life, Physical, & Social Science, Entertainment, Sports,
https://x.com/arena/status/2034653740465336407

With the preview of Qwen 3.5 Max Preview by @Alibaba_Qwen, we’re looking back at past Qwen Max variants to see how far it has progressed. Where Qwen 3.5 Max sees the largest gains vs. Qwen 3 Max: – Text Overall (+45pts) – Creative Writing (+57pts) – Math (+49pts) –
https://x.com/arena/status/2034658045113065603

ScreenSpot-Pro, the GUI computer use benchmark is now on @huggingface 🏆 just added Qwen3.5 it takes 5th place, with specialist Holo2 family takes top ranks whoever builds next GUI model based on Qwen3.5 can top the leaderboard? 🔥
https://x.com/mervenoyann/status/2034265145158119642#m

Good news: I got Qwen3.5-397B-FP8 running on my 8x mi210 server. Bad news: at 6 tokens per second.
https://x.com/QuixiAI/status/2033342155414982952

We’re re-releasing Open SWE with a new emphasis on deep integrations with the applications you already use It integrates with: – Slack – Linear – Github so you can use it directly from the applications you already spend your time in, without needing to learn a new platform
https://x.com/BraceSproul/status/2033962118970818650

Leave a Reply

Trending

Discover more from Ethan B. Holland

Subscribe now to keep reading and get access to the full archive.

Continue reading