Image created with gemini-2.5-flash-image with claude-sonnet-4-5. Image prompt: Photorealistic 35mm cinema shot of child viewing TV screens in warm bedroom, shallow depth of field with robotic arm arranging books on shelf in background, small autonomous device on plush rug edge, screens displaying UI fragments and task notifications, warm peach lighting contrasted with cool blue screen glow, side angle composition, cozy domestic automation, large bold text AGENTS at top
Anthropic preparing new Agentic Tasks Mode for Claude https://www.testingcatalog.com/anthropic-testing-new-agentic-tasks-mode-for-claude/
Fastweb + Vodafone (Swisscom Group), one of Europe’s leading telecom providers, is building Super TOBi, which brings agentic customer service to massive scale. Using LangSmith, they are: 🔹Achieving 90% response correctness and 82% resolution rates across ~9.5M customers https://x.com/LangChain/status/2001321491703443877
GROK JUST TURNED VOICE AI INTO A REAL PRODUCT, FAST, AND EVERYWHERE xAI just opened Grok Voice to developers, and this isn’t some early experiment dressed up as a launch. It’s the same system already running inside millions of Teslas, now exposed through an API that actually https://x.com/MarioNawfal/status/2001472484869329288
Grok Voice Agent API | xAI https://x.ai/news/grok-voice-agent-api
Today, we’re excited to launch the Grok Voice Agent API, empowering developers to build voice agents that speak dozens of languages, call tools, and search realtime data. https://x.com/xai/status/2001385958147752255
Took less than an hour for Grok Voice Agent by @xai to be ported to Reachy Mini thanks to @atariorbit! https://x.com/ClementDelangue/status/2001410494528213481
Sonnet 4.5 was underestimated on METR its time horizon improves around 20 minutes https://x.com/scaling01/status/2001476927362605354
We’re working on updating and improving our time horizon task suite. Recently, we found two issues with our tasks, one of which was differentially lowering the performance of Claude models. We think these also illustrate some interesting model behavior.”” / X https://x.com/METR_Evals/status/2001473506442375645
IBM dropped CUGA, open-source enterprise agent to automate boring tasks 🔥 > given workspace files, it writes and executes code to accomplish any task 🤯 > comes with a ton of tools built for enterprise tasks, supports MCPs > plug in your favorite LLM 👏 here’s a small demo https://x.com/mervenoyann/status/2000599316121924052
Skills Directory | Partner Skills for Claude – YouTube
Skills for organizations, partners, the ecosystem | Claude https://claude.com/blog/organization-skills-and-directory
Gemini Agent can help tackle all sorts of tasks. Even renting a car. Tell Gemini Agent your budget and it’ll get to work comparing prices, gathering info from your inbox, and booking the car. Now available for Google AI Ultra users in the US on desktop and mobile.”” / X https://x.com/GeminiApp/status/2000616120106221781
Gemini 3, create a really novel and clever and funny Venn diagram. think hard. do not do research.”” So close to coming together (I am not sure the center works for all three, illustrations are odd), but also better than I expected. https://x.com/emollick/status/2000805347590856822
Gemini 3, please provide the rail/subway map for Middle Earth in the third age, with accurate stops and taking into account natural barriers, alliances, and so on.”” Not bad. I do like the “”service suspended – Balrog”” note at Moria. https://x.com/emollick/status/1999930443001737700
Gemini can now illustrate a visual report https://blog.google/products-and-platforms/products/gemini/visual-reports/
Google Antigravity https://antigravity.google/
Google expands Gemini with NotebookLM integration https://www.testingcatalog.com/google-expands-gemini-with-notebooklm-integration/
Build mini apps with Opal in the Gemini web app https://blog.google/innovation-and-ai/models-and-research/google-labs/mini-apps-opal-gemini-app-experiment/
Say hello to CC, a new AI productivity agent that connects your Gmail, Calendar and Drive to deliver a personalized briefing every morning. Need more help? Just email CC https://labs.google/cc
After a day of gemini 3 flash in antigravity, I think I’m convinced. It’s really good to have a lightning fast and smart model for daily work. I’ve been pretty adamant that slower is ok if the model is smarter, but the models have produced just slightly too much cruft and I”” / X https://x.com/andrew_n_carr/status/2001487412749570549
For a fast model, Gemini 3 Flash offers incredible performance, allowing us to provide frontier intelligence to everyone globally. Try the ‘fast’ mode from the model picker in the @GeminiApp – it’s shockingly speedy AND smart. Best pound-for-pound model out there ⚡️⚡️⚡️ https://x.com/demishassabis/status/2001325072343306345
For developers, it combines advanced coding skills with the low latency needed for building interactive apps. On SWE-bench Verified – a benchmark for evaluating coding agents – it outperforms not only the 2.5 series, but also Gemini 3 Pro. Watch 3 Flash give near real-time AI https://x.com/GoogleDeepMind/status/2001321765503377546
Gemini 3 Flash gives you frontier intelligence at a fraction of the cost. ⚡ Here’s how it’s built for speed and scale 🧵 https://x.com/GoogleDeepMind/status/2001321759702663544
Gemini 3 flash is a bigger deal than Gemini 3 pro. While 2.5 flash is the most used model this year, but it struggled with tool calling. But Gemini 3 flash gets it. – tool calling feels natural to the model – it’s faster than turbo models + way smarter too (best for real time”” / X https://x.com/0xdevshah/status/2001330346961604732
Gemini 3 Flash is beating 3 Pro on SWE bench verified Hmm what https://x.com/MS_BASE44/status/2001698991801798927
Gemini 3 Flash is starting to roll out in the @GeminiApp and across Google products. Learn more ↓ https://x.com/Google/status/2001746491275083925
Gemini 3 Flash punches way above its weight class, surpassing 2.5 Pro on many benchmarks, while being much cheaper, faster, and more token efficient. https://x.com/OfficialLoganK/status/2001323840459456715
Google has released Gemini 3 Flash Preview – 2x cheaper than Gemini 3 Pro Preview, with only a 2-point drop in our Intelligence Index, making it the most intelligent model for its price range @GoogleDeepMind gave us pre-release access to Gemini 3 Flash Preview. The model scores https://x.com/ArtificialAnlys/status/2001335953290670301
Gemini 3.0 Flash is an absolutely fantastic release. Consider this: It costs a quarter (1/4) of what Gemini 3.0 Pro costs and achieves similar results to the Pro model in almost all benchmarks, such as HLE and ARC-AGI 2. In other benchmarks, it even outperforms the more https://x.com/kimmonismus/status/2001326181875154983
Introducing Gemini 3 Flash: Benchmarks, global availability https://blog.google/products-and-platforms/products/gemini/gemini-3-flash/
Starting today, Gemini can serve up local results in a rich, visual format. See photos, ratings, and real-world info from @GoogleMaps, right where you need them.”” / X https://x.com/GeminiApp/status/1999631529379791121
BREAKING: OpenAI releases “”GPT-Image-1.5″” (ChatGPT Images) & It instantly takes the #1 Spot on LMArena, beating Google’s Nano Banana Pro. : r/singularity https://www.reddit.com/r/singularity/comments/1po98xo/breaking_openai_releases_gptimage15_chatgpt/
A year ago, we verified a preview of an unreleased version of @OpenAI o3 (High) that scored 88% on ARC-AGI-1 at est. $4.5k/task Today, we’ve verified a new GPT-5.2 Pro (X-High) SOTA score of 90.5% at $11.64/task This represents a ~390X efficiency improvement in one year https://x.com/arcprize/status/1999182732845547795
GPT-5.2-Codex launches today. It is trained specifically for agentic coding and terminal use, and people at OpenAI have been having great success with it.”” / X https://x.com/sama/status/2001724019188408352
Meet GPT-5.2-Codex, the best agentic coding model yet for complex, real-world software engineering. With native compaction, better long-context understanding, and improved tool-calling, it is a more dependable partner for your hardest tasks. Available in Codex starting today.”” / X https://x.com/OpenAIDevs/status/2001723687373017313
GPT-5.2 exceeded a trillion tokens in the API on its first day of availability and is growing fast!”” / X https://x.com/sama/status/1999624463013544024
I have found GPT-5.2 Thinking to be a surprisingly deep second-opinion/fact checker. I gave it a dense paragraph with a few correct claims, a couple errors that required research to find, and some things that needed interpretation It found and gently corrected all the problems https://x.com/emollick/status/2000666007010971787
Introducing GPT-5.2-Codex | OpenAI https://openai.com/index/introducing-gpt-5-2-codex/
GPT-5.2 is here and it’s the best model out there for everyday professional work. On GDPval, the thinking model beats or ties human experts on 70.9% of common professional tasks like spreadsheets, presentations, and document creation. It’s also better at general intelligence,”” / X https://x.com/fidjissimo/status/1999183159356006450
Today I ran two complex tasks through Codex with GPT 5.2 Extra High The first ran for 2 hours 30 minutes The second ran for 1 hours 45 minutes Both resulted in: – all acceptance criteria resolved – all test coverage complete – zero broken or non-working code Amazing”” / X https://x.com/nummanali/status/2000228337030152347
Whoa. This new GDPval score is a very big deal. Probably the most economically relevant measure of AI ability suggesting that in head-to-head competition with human experts on tasks that require 4-8 hours for a human to do, GPT-5.2 wins 71% of the time as judged by other humans https://x.com/emollick/status/1999189828756263359
xAI’s new Grok Voice Agent is the new leading Speech to Speech model, surpassing Gemini 2.5 Flash Native Audio and GPT Realtime in our Big Bench Audio benchmark The new model achieves a score of 92.3% on Big Bench Audio, just ahead of the previous leader, Google’s Gemini 2.5 https://x.com/ArtificialAnlys/status/2001388724987527353
How good is AI for science? Yesterday, OpenAI released a benchmark, FrontierScience, to measure frontier model performance on scientific tasks. This is the most sophisticated benchmark for science I’ve seen. FrontierScience has 160 questions across various subdomains, https://x.com/jungofthewon/status/2001302379527114798
⚖️ Pairwise Annotations: Scores are hard, preferences are easy. Agents handle tasks that are tough to score but easy to compare: support responses where tone matters, code refactors where both work but one feels cleaner, product specs where “”good”” is subjective. In practice, https://x.com/LangChain/status/2001361753851203724
When Machines Pay Machines: The Economics of Agentic AI | Matt Suiche https://www.msuiche.com/posts/when-machines-pay-machines-the-economics-of-agentic-ai/
@swyx To me agents and harnesses are fully coupled, not really possible to properly eval one without the other. Currently workshopping: if agent = folder, what goes in there, how do we install/assemble agents? So here’s current mental model of both agents & harnesses with some”” / X https://x.com/Vtrivedy10/status/2001868118952436103
⚡ Faster than Fast. Designed for Agentic AI. Introducing Xiaomi MiMo-V2-Flash — our new open-source MoE model: 309B total params, 15B active. Blazing speed meets frontier performance. 🔥 Highlights: 🏗️ Hybrid Attention: 5:1 interleaved 128-window SWA + Global | 256K context 📈 https://x.com/XiaomiMiMo/status/2000929154670157939
🔌 New in LangChain MCP Adapters (feat 3/4): structured content from tools 📦 MCP tools can now return content and structuredContent (often JSON payloads and pydantic models) –ideal for agents exposed as MCP tools! Docs: https://x.com/sydneyrunkle/status/1999538200243511725
🚀 Deep Agents: The Weekly Roundup 🚀 Dive into our latest resources to help you build, observe, and evaluate Deep Agents capable of handling complex, long-running tasks. 📊 How to Observe Deep Agents – Agents are running longer and getting more complex, which demands new https://x.com/LangChainAI/status/1999568074450829482
🚀 We just launched OpenHands Software Agent SDK on @ProductHunt! A smarter way to build agent-driven software — fast, flexible, and production-ready. 👉 Check it out + show some love! https://x.com/OpenHandsDev/status/2000805627967209728
2026 vibe coding tool comparison – by Justin – Technically https://read.technically.dev/p/2026-vibe-coding-tool-comparison
6 Comprehensive resources on AI Coding ▪️ AI Agentic Programming: A Survey of Techniques, Challenges, and Opportunities ▪️ Does AI-Assisted Coding Deliver? A Difference-in-Differences Study of Cursor’s Impact on Software Projects ▪️ A Survey of Vibe Coding with LLMs ▪️ From Code https://x.com/TheTuringPost/status/2000171190506373336
6 most populat Policy Optimization algorithms in 2025 ▪️ PPO (Proximal Policy Optimization) ▪️ GRPO (Group Relative) ▪️ GSPO (Group Sequence) ▪️ DAPO (Decoupled Clip and Dynamic sAmpling) ▪️ BAPO (BAlanced) ▪️ ARPO (Agentic Reinforced) Learn more about each method and the key https://x.com/TheTuringPost/status/1999801691538104543
Agent Skills are now supported in Stirrup – our lightweight framework for building agents Using Agent Skills in Stirrup is as easy as specifying the directory of your skills files (typically just markdown files). Agent Skills are folders of instructions, scripts, and resources https://x.com/ArtificialAnlys/status/2001778418590060819
Agent Skills is now an open standard It’s been great to see the traction Skills are already getting in the industry and this makes it easier for everyone to build and contribute to them🚀 https://x.com/alexalbert__/status/2001760879302553906
DeepCode – a multi-agent framework that turns research papers into full codebases It manages information flow so large, detailed papers can be converted into production-quality code despite LLM context limits. It does this through: – Blueprint distillation – compressing papers https://x.com/TheTuringPost/status/1999781163976843282
Do you want to run coding agents safely, without damaging to your filesystem? 📁 Last week, we published a blog post and a demo showing exactly how to do this with @claudeai and AgentFS by @tursodatabase. After strong community interest, we’ve now shipped support for @OpenAI https://x.com/llama_index/status/2002064702927769706
Graphite is joining Cursor · Cursor https://cursor.com/blog/graphite
I’m bullish on @temporalio. Their abstractions are strong in general but are a particularly good match for long-running background agents.”” / X https://x.com/corbtt/status/2001801936916643919
Klarna launches Agentic Product Protocol: The open standard that makes 100M+ products instantly discoverable by AI agents | Klarna International https://www.klarna.com/international/press/klarna-launches-agentic-product-protocol-the-open-standard-that-makes-100m/
Overview – Agent Skills https://agentskills.io/home
Prompt caching: 10x cheaper LLM tokens, but how? | ngrok blog https://ngrok.com/blog/prompt-caching
Secure your coding agents with virtual filesystems and better document understanding. Building safe AI coding agents requires solving two critical challenges: filesystem access control and handling unstructured documents. We’ve created a solution using AgentFS, LlamaParse, and https://x.com/llama_index/status/2000612235505467824
The Cline provider now runs on the @vercel AI Gateway. The move delivers immediate, measurable improvements. Error rates are down 43.8% (from 1.78% to 1%); production testing shows P99 streaming latencies improved 10-40% across our most popular models. Grok-code-fast-1 saw P99 https://x.com/cline/status/2001043584490070470
Tired of waiting for minutes for your AI coding assistant? @cognition built agents that search, reason & edit code in a few seconds. Powered by Cerebras–running at 1K tokens/sec with frontier-level accuracy. https://x.com/cerebras/status/1999540379553611955
Towards a Science of Scaling Agent Systems https://arxiv.org/pdf/2512.08296
We’re excited to share that LangSmith, the agent engineering platform for observability, evaluation, and deployment, has been named one of @brexHQ’s Top 25 Fastest-Growing Software Vendors of 2025 🎉”” / X https://x.com/LangChain/status/2001321495037985194
We’re hitting a massive wall in the AI stack: serverless backends are choking on agents. Developers are hitting execution ceilings and resorting to network hacks just to keep their apps alive and responsive. Reliable agentic loops require persistent, long-running infrastructure”” / X https://x.com/anuraggoel/status/2001721861198221629
what’s actually in an agent & harness? how do we engineer them? what are coding agent products bundling into their harness to get good performance & UX? seeing a lot of these questions so did a pass at re-distilling my mental model in this tweet response the TLDR: “”” Agent =”” / X https://x.com/Vtrivedy10/status/2002077611548135756
When agents help us write an order of magnitude more code, the bottleneck for software engineering is review.”” / X https://x.com/amanrsanger/status/2002090644127560085
working through mental model of are the blocks that go in an agent + terminology: – agents are a box/folder that we put composable blocks inside – right now these blocks should just be prompts, skills, subagents (optional), memory (optional) -> skill pilled – but every block”” / X https://x.com/Vtrivedy10/status/2001682603460473190
Replit — Inside Replit’s Snapshot Engine: The Tech Making AI Agents Safe https://blog.replit.com/inside-replits-snapshot-engine
When Agents Attack: How AI Collapses and Rebuilds Marketplace Moats https://www.caseyaccidental.com/p/when-agents-attack-how-ai-collapses
Claude Code 🤝 LangSmith Curious what Claude Code is doing behind the scenes? Or want observability into critical workflows that you’ve set up with Claude Code. With our new Claude Code → LangSmith integration, you can view every 🤖 LLM call and 🔧 tool call Claude Code makes. https://x.com/LangChain/status/2002055677708058833
i was skeptical when @simonw said that “”Claude Skills are awesome, maybe a bigger deal than MCP”” buuut early indications are this is correct. this is the fastest talk ever to pass 100k views here at AIE. its like those 0 – 100m ARR charts but for attention. @MaheshMurag and https://x.com/swyx/status/1998786773477110049
LangSmith + Claude Code / Deepagents Pairing LangSmith tracing w/ code agents provides a powerful feedback loop. Here, we show examples of that w/ langsmith-fetch + Claude Code / Deepagents. langsmith-fetch CLI: https://x.com/LangChain/status/2001350950188126430
The Signature Flicker | Peter Steinberger https://steipete.me/posts/2025/signature-flicker
We now support Agent Skills – the open standard created by @AnthropicAI for extending AI agents with specialized capabilities. Create skills once, use them everywhere. 🔗 https://x.com/code/status/2001727543377039647
vibe coding games is actually a lot of fun. can’t wait to share something cool soon.”” / X https://x.com/bilawalsidhu/status/1998961420457881654
Ad: Pretty cool to vibe code games using YouTube Playables Builder. One of my top VFX/360 videos is now a retro shooter game – stock up on burgers for your intergalactic overlords while dodging a horde of farmers who really want their cows back. https://x.com/bilawalsidhu/status/2001025884778848611
Measuring AI Ability to Complete Long Tasks – METR https://metr.org/blog/2025-03-19-measuring-ai-ability-to-complete-long-tasks/
The updated time horizon numbers are live on the dashboard on our website: https://x.com/METR_Evals/status/2001473519197335899
Best AI research of the week: ▪️ On the Interplay of Pre-Training, Mid-Training, and RL on Reasoning LMs ▪️ Native Parallel Reasoner ▪️ Long-horizon Reasoning Agent for Olympiad-Level Mathematical Problem Solving ▪️ DoVer: Intervention-Driven Auto Debugging for LLM Multi-Agent https://x.com/TheTuringPost/status/2000874193249034463
AI agents are starting to eat SaaS – Martin Alderson https://martinalderson.com/posts/ai-agents-are-starting-to-eat-saas/
Lovable raises $330M to power the age of the builder – Lovable Blog https://lovable.dev/blog/series-b
Got to talk about one of my most contrarian takes with @PeterDiamandis: I think the AGI “”race”” is a huge misnomer. There won’t be one single AGI, but infinite variations and styles and forms. And that’s way more exciting than a zero-sum race. https://x.com/mustafasuleyman/status/2001374004960203048
DeepCode: Open Agentic Coding AI coding agents still can’t reliably turn research papers into working code. The best LLM agents achieve only 42% replication scores on scientific papers, while human PhD experts hit 72%. But the problem isn’t model capability. This new paper https://x.com/omarsar0/status/2000385348413850055
The Adoption and Usage of AI Agents: Early Evidence from Perplexity https://arxiv.org/pdf/2512.07828
Chain of Unit-Physics builds physics knowledge directly into the code generation process. Researchers from @UMich propose an inverse approach to scientific code generation: – They encode human expert knowledge as unit-physics tests that the code must pass. – In a multi-agent https://x.com/TheTuringPost/status/2000177305981944308
InternGeometry: An LLM Agent tackles Olympia-level geometry. This novel agent solves 44 of 50 International Math Olympiad problems, beating gold medalists with only 13K training examples. It uses iterative reasoning & Complexity-Boosting RL. https://x.com/HuggingPapers/status/1999572332906438987
A thing that the other models need to copy from Claude is a switch that lets you turn off web search. Now that all the models are good at using tools, they turn to the web too often when sometimes you just want the model to take what you put in the context window & work with that https://x.com/emollick/status/2000807086880694752
Claude Skills can accomplish a lot of hard tasks & are accessible to non-technical people, but hidden behind a somewhat intimidating technical gloss. With some better user experience, they are a natural sequel to GPTs as a way for people inside organizations to innovate with AI.”” / X https://x.com/emollick/status/1999148820668555520
First Look: Unboxing Guardrails for AI-Generated Code https://webinars.sonatype.com/wcc/eh/5011667/lp/5151488/first-look-unboxing-guardrails-for-ai-generated-code/
harnesses are distribution mechanisms for good tooling and taste each choice helps craft the ✨experience✨ for the user planning view, context management on behalf of user, specialized subagents we think are useful, UX flow for viewing subagents, memory updates UX, parallel”” / X https://x.com/Vtrivedy10/status/2001492640076894661
Interpretability agents are a big deal for researchers. But they’re a pain – research is so custom! Seer has many quality of life improvements to make research with agents easy. It’s hackable & extensible, to enable as much research as possible, incl weird cursed techniques!”” / X https://x.com/NeelNanda5/status/2002051650949943346
Official rule for all AI labs: no more demoing your product with either telling the AI to “book a trip for me” or creating AI photos/videos of your company’s CEO in crazy situations. Sorry, those are the rules now. https://x.com/emollick/status/2001119366557900914
What Actually Is Claude Code’s Plan Mode? | Armin Ronacher’s Thoughts and Writings https://lucumr.pocoo.org/2025/12/17/what-is-plan-mode/
Project Vend: Phase two \ Anthropic https://www.anthropic.com/research/project-vend-2
Jeff and Sanjay’s work at Google which includes code commits and performance optimization is probably something only ASI can match!! So happy to have access to this document which is soo goood! I feel like this performance oriented thinking rubbed off on many folks at the”” / X https://x.com/_arohan_/status/2002105340062552509
Say hello to the new Interactions API and our first agent, Gemini Deep Research, now available for developers 🤖! The Interactions API is a new unified interface to interact with both models and agents. Our Deep Research agent is also SOTA on many dimensions…”” / X https://x.com/OfficialLoganK/status/1999163355525956020
We still don’t exactly know when agents help and when they hurt. Many design them by intuition. Google outlined practical principles for how agent systems scale: – More agents is not always better – Strong single agents don’t benefit much from coordination – Coordination has https://x.com/TheTuringPost/status/1999499042880127328
An engineer showed Gemini what another AI said about its code Gemini responded (in its “”private”” thoughts) with petty trash-talking, jealousy, and a full-on revenge plan 🧵 https://x.com/AISafetyMemes/status/2000620127054598508
We spun up a new GitHub repo for all things MCP at @Google. Get info on our remote managed MCP servers, open source MCP servers, examples, and learning resources. https://x.com/rseroter/status/2000607267675410609
New benchmark from Google Research. Models get better at benchmarks, but do they actually get more factual? Previous evaluations focused on narrow slices: grounding to documents, answering from memory, or using search. A model excelling at one often fails at another. This new https://x.com/omarsar0/status/2000935220049273303
after testing GPT-5.2 I no longer think that it is a much larger model or anywhere the size Gemini 3 Pro is”” / X https://x.com/scaling01/status/1999566015873569174
Our team at Google DeepMind is hiring iOS engineers 🙂 come build the future of vibe coding with us! https://x.com/OfficialLoganK/status/2000662065074131221
Google names new chief of AI infrastructure buildout | Semafor https://www.semafor.com/article/12/10/2025/google-names-new-chief-of-ai-infrastructure-buildout
Google’s AI Playbook for Sustainability Reporting https://blog.google/company-news/outreach-and-initiatives/sustainability/ai-playbook-sustainability-reporting/
Gemini 3 Pro continues to be SOTA at multimodal understanding and generation : ) cc @bcaine for the great example https://x.com/OfficialLoganK/status/1999270402712023158
Gemini 3 Pro playing Pokémon vs 2.5 Pro (we used to all be impressed by 2.5 Pro) https://x.com/OfficialLoganK/status/2000728193599226187
Google Translate gets new Gemini AI translation models https://blog.google/products-and-platforms/products/search/gemini-capabilities-translation-upgrades/
🌎 Google’s FunctionGemma is 270M model that’s fine-tuned by Google for function calling. Try it on Ollama’s latest v0.13.5 ollama pull functiongemma examples on model page 👇👇👇 https://x.com/ollama/status/2001705006450565424
Fine tune Google’s FunctionGemma for Mobile, with agents, on colab, locally, or Hugging Face. Google Deepmind Have just release FunctionGemma and anyone can finetune it with TRL. This is the model: – uses the Gemma 3 270M architecture + adapted chat template – specifically for https://x.com/ben_burtenshaw/status/2001704049490489347
FunctionGemma – a google Collection https://huggingface.co/collections/google/functiongemma
Google is preparing for a new open source release on @huggingface Also noticed just recently that Gemma models are not available on AI Studio anymore. What do you expect? 👀 https://x.com/testingcatalog/status/2000597370707611991
I’m very excited to release Gemma Scope 2: Sparse Autoencoders, and transcoders on every layer of every Gemma 3 model: 270M to 27B, base and chat We want to make it easier to do deep dives into interesting model behaviour, I’m excited to see what you all can do with them”” / X https://x.com/NeelNanda5/status/2002080911693643806
Introducing FunctionGemma 🤏270m model for function calling 📱can run in your phone, browser or other devices 🤖designed to be specialized for your own tasks https://x.com/osanseviero/status/2001704034667769978
Introducing T5Gemma 2, the next generation of encoder-decoder models 🚀 Built on top of Gemma 3, we were able to build compact models at sizes of 270m-270m, 1B-1B, and 4B-4B sizes. While most models today are decoder-only, T5Gemma 2 is the first (I’m aware of) multimodal, https://x.com/osanseviero/status/2001723652635541566
To build safer AI, we need to understand how models “”think””. 🧠 Enter Gemma Scope 2, a new set of tools to interpret Gemma 3: our family of lightweight open models. It can help researchers trace internal reasoning, debug complex behaviors and identify risks → https://x.com/GoogleDeepMind/status/2002018669879038433
Update: Gemma 4 incoming! Let’s go, google!”” / X https://x.com/kimmonismus/status/2000537345326452790
We made 3 @UnslothAI tool calling notebooks for FunctionGemma! 1. Fine-tuning it to make it reason before tool calling 2. Multi-turn tool calling 3. Tool calling fine-tuning to enable mobile actions Guide: https://x.com/danielhanchen/status/2001713676747968906
e
🚨BREAKING: Leaderboard updates for Text, Vision & WebDev Gemini-3-Flash by @GoogleDeepMind is now ranked top 5 across Text, Vision, and WebDev, making it the most cost-efficient frontier model (input $0.5 and output $3/MTokens). Gemini-3-Flash highlights: 🔹 Top 5 across Text, https://x.com/arena/status/2001322123730788698
📢 New Model(s) Drop: Gemini 3 Flash Preview is now live on Yupp’ The latest from @GoogleDeepMind offers frontier-level intelligence with reduced costs and more speed. Ready to test it out? It’s available on Yupp in several variants! https://x.com/yupp_ai/status/2001340530828206586
Gemini 3 Flash above GPT-5.2 on EpochAI’s ECI https://x.com/scaling01/status/2001850867620946169
Gemini 3 Flash is now available ⚡ Since introducing the Gemini 3 series last month, we’ve seen you vibe code simulations to learn about complex topics, build and design interactive websites and understand multimodal content. Now we’re introducing Gemini 3 Flash, our latest https://x.com/Google/status/2001322381533409733
Gemini 3 Flash is now available in Cursor! We’ve found it to work well for quickly investigating bugs.”” / X https://x.com/cursor_ai/status/2001326908030804293
Gemini 3 Flash is now available to all Perplexity Pro and Max subscribers. https://x.com/perplexity_ai/status/2001447398317724153
Gemini 3 Flash is now rolling out to @code developers! https://x.com/pierceboggan/status/2001327058425917795
Gemini 3 Flash is rolling out globally today. ⚡⚡⚡ Let us know how you’re using it in the replies ↓”” / X https://x.com/GeminiApp/status/2001412101286563865
Gemini 3 Flash is the new default for vibe coding”” / X https://x.com/OfficialLoganK/status/2001352972379549721
Gemini 3 Flash Low on LisanBench – low does obviously worse than high – still inefficient reasoning, ~2x lower score for ~2x less tokens – validity ratios are absolutely abysmal https://x.com/scaling01/status/2001359254578753852
Gemini 3 Flash on the @ArtificialAnlys intelligence benchmark, the most cost per intelligence efficient model in the world!!! https://x.com/officiallogank/status/2001368440016392314
Gemini 3 Flash on the @ArtificialAnlys intelligence benchmark, the most cost per intelligence efficient model in the world!!! https://x.com/OfficialLoganK/status/2001368440016392314
Gemini 3 Flash Preview ranking 5th on SimpleBench ahead of GPT-5.2 Pro https://x.com/scaling01/status/2002024316842512812
Gemini 3 Flash ranks #3 in the LMArena leaderboard (which is especially notable given its API pricing and its low latency).”” / X https://x.com/JeffDean/status/2001335803642024157
Gemini 3 Flash rolling out to @code now 🚀 Try it out and let us know what you think! https://x.com/code/status/2001335940934246503
Gemini 3 Flash scores higher than GPT-5.2, Opus 4.5 and Gemini 3 Pro on SWE-Bench Verified ??? https://x.com/scaling01/status/2001803023811797433
Gemini 3 Flash takes the #1 spot on Toolathlon https://x.com/scaling01/status/2001849103647674538
Gemini 3.0 Flash achieved a very impressive 161.8/190 on one of my vibe tests, the Korean Sator Square Test (KSST), placing it 2nd or 3rd among all the models I have tested so far. This is slightly higher than Gemini 3.0 Pro, and the difference is within the margin of error. https://x.com/Hangsiin/status/2001341564145250770
Going live with the team in a few to talk about Gemini 3 Flash : ) send us your questions! https://x.com/OfficialLoganK/status/2001372183663378723
How good is Gemini 3 flash? “”We ran a behind-the-scenes test with 3 Flash. Because of how much faster it was, retention went up, the number of things people were building went up, and engagement went up.”””” / X https://x.com/_philschmid/status/2001492609114456471
🗣️ “”Help me build an app…”” That’s all it takes. Watch Gemini 3 Flash turn a single voice prompt into a functional prototype in the @GeminiApp. https://x.com/Google/status/2002123256854425918
Congrats to the Gemini team on the great release and exceptional SWE-bench Verified numbers! 76.2% (3 Pro) vs. 78% (3 Flash), +6 task instances – a whole lot in the realm of the last quarter of SWE-bench. mini-SWE-agent + Gemini 3 Flash coming soon!”” / X https://x.com/jyangballin/status/2001336879120363639
Gemini 3 Flash across different test-time compute levels (green line below) represents a new score/cost Pareto frontier on ARC-AGI-2. Congrats to @demishassabis and @sundarpichai on the launch! https://x.com/fchollet/status/2001330643423449409
Gemini 3 Flash is out ⚡️- and we built a CLI agent powered by this latest model to perform work over your filesystem 🤖 Basically all the file capabilities within Claude Code in a lighter form factor. Shoutout to @itsclelia for the launch demo, check it out! Repo: https://x.com/jerryjliu0/status/2001335494534402521
how can flash beat pro??”” -> the answer is RL! flash is not just a distilled pro. we’ve had lots of exciting research progress on agentic RL which made its way into flash but was too late for pro. can’t wait to finally bring them to pro👀”” / X https://x.com/ankesh_anand/status/2002017859443233017
Introducing Gemini 3 Flash ⚡️Performance close to Gemini 3 Pro, with great multimodal and tool use quality ⚡️3x faster than Gemini 2.5 Pro, while cheaper and better at most benchmarks ⚡️LMArena score of 1477 (top 3 model) The time to build is now (and yes, there’s a free tier)”” / X https://x.com/osanseviero/status/2001323721232163053
Introducing Gemini 3 Flash, our frontier intelligence model, available at scale for everyone. It excels at coding, tool calling, and is stronger than 2.5 Pro across most metrics!! ⚡️ Available in the API at $0.50 in / 1M tokens and $3.00 out / 1M tokens across. https://x.com/OfficialLoganK/status/2001322275656835348
Introducing Gemini 3 Flash! ⚡️⚡️⚡️ Frontier intelligence built for speed at a fraction of the cost. Here’s ~4 minutes of demos. https://x.com/addyosmani/status/2001324727504359745
Speed test: Gemini 3 Flash vs. Gemini 2.5 Pro ⏱️ We put our new Gemini 3 Flash model (left) up against Gemini 2.5 Pro (right) in @GoogleAIStudio, so you can watch the difference in near real-time. Watch them go head-to-head ↓ https://x.com/Google/status/2001397324551946523
Study with help from Gemini 3 Flash. Upload an audio recording of yourself explaining a difficult concept and Gemini will identify knowledge gaps, create a custom quiz, and provide instant assessments and explanations for each question.”” / X https://x.com/GeminiApp/status/2001351746338329063
Today, we’re releasing an updated Gemini 2.5 Flash Native Audio model. Now available via the Live API 🗣 https://x.com/googleaidevs/status/1999539531826036973
Watch Gemini 3 Flash vs Gemini 3 Pro playing Pokemon Crystal : ) https://x.com/OfficialLoganK/status/2001428651121025391
We’re back in a Flash ⚡ Gemini 3 Flash is our latest model with frontier intelligence built for lightning speed, and pushing the Pareto Frontier of performance and efficiency. It outperforms 2.5 Pro while being 3x faster at a fraction of the cost. With this release, Gemini 3’s https://x.com/sundarpichai/status/2001326061787942957
We’re expanding the Gemini 3 family with the launch of Gemini 3 Flash. This model: — Combines Gemini 3’s Pro-grade reasoning with Flash-level latency, efficiency, and cost — Delivers frontier-level performance on PHD-level reasoning and knowledge benchmarks — Is our most https://x.com/googleai/status/2001323069105692914
we’re going live at 11:30am PT with the team for a deep dive on gemini 3 flash hosted by @OfficialLoganK, @joshwoodward, @tulseedoshi and more post your questions below ⬇️ https://x.com/GoogleAIStudio/status/2001330099841556490
We’ve pushed out the Pareto frontier of efficiency vs. intelligence again. With Gemini 3 Flash ⚡️, we are seeing reasoning capabilities previously reserved for our largest models, now running at Flash-level latency. This opens up entirely new categories of near real-time https://x.com/JeffDean/status/2001323132821569749
With Gemini 3 Flash, you can quickly build fun, useful apps from scratch using your voice without any prior coding knowledge. Just dictate to Gemini on the go, and it can transform your unstructured thoughts into a functioning app in minutes.”” / X https://x.com/GeminiApp/status/2001760080518353261
Realtime speech to speech translation powered by Gemini, available in Google Translate now, coming to developers early next year : ) https://x.com/OfficialLoganK/status/1999994009452962073
Nvidia and Alphabet VC arms back vibe coding startup Lovable https://www.cnbc.com/2025/12/18/google-and-n.html
NEW: Google releases FunctionGemma, a lightweight (270M), open foundation model built for creating specialized function calling models! 🤯 To test it out, I built a small game: use natural language to solve fun physics simulation puzzles, running 100% locally in your browser! 🕹️ https://x.com/xenovacom/status/2001703932968452365
Ever opened a repo and thought: “What does this codebase actually do?” “Where did I put that file?” 🤔 You’re not alone. With the release of Gemini 3 Flash ⚡ from @GoogleDeepMind, we decided to build something fun (and useful): a file-system explorer agent that answers those https://x.com/llama_index/status/2001324278617424017
FunctionGemma has day-0 support on MLX 🔥🚀 A tiny but mighty single-turn function calling model. Great for on-device tool use, MCP, RAG, routing and more. Get started today: > pip install -U mlx-lm Or run it on your iPhone using MLX-Swift. Notebook example: https://x.com/Prince_Canuma/status/2001713991115026738
Opera launches Neon browser globally with paid early access https://www.testingcatalog.com/icymi-opera-launches-neon-browser-globally/
Local, cloud, and background agents, all in a unified experience in @code https://x.com/code/status/1999575448087396563
(13) Don’t Build Agents, Build Skills Instead – Barry Zhang & Mahesh Murag, Anthropic – YouTube https://www.youtube.com/watch?v=CEvIs9y1uog
An implicit belief at the Ai labs is that spending too much time productizing around the weaknesses of current models is a waste because better models will solve many of those issues. They may be right: Copilot was built well to address the gaps from GPT-4, but now has to pivot.”” / X https://x.com/emollick/status/2001030095826510184
Copilot just got smarter! Starting today, we’re rolling out the latest GPT-5.2 model from our partners at OpenAI to consumer @Copilot, coming first to Microsoft 365 Premium users. Can’t wait to see what you do with it.”” / X https://x.com/mustafasuleyman/status/1999184598987866194
.@MistralAI’s Devstral 2 family of models are now available in Ollama. 24B: ollama run devstral-small-2 123B: ollama run devstral-2 Ollama’s cloud: ollama run devstral-2:123b-cloud https://x.com/ollama/status/1999590723373662612
codex now supports skills, per the https://x.com/gdb/status/2002120466203615649
GPT-5.2-Codex is now available in Codex. It sets a new standard for agentic coding in real-world software development and defensive cybersecurity. It also delivers more reliable performance on complex tasks and scales effectively across large projects. https://x.com/OpenAI/status/2001766212494332013
Have Codex automatically fix GitHub CI failures $.skill-installer gh-fix-ci https://x.com/OpenAIDevs/status/2002100589732508010
Have Codex read and update your Linear tickets. $.skill-installer linear https://x.com/OpenAIDevs/status/2002099775634878930
🆕 Codex now officially supports skills Skills are reusable bundles of instructions, scripts, and resources that help Codex complete specific tasks. You can call a skill directly with $.skill-name, or let Codex choose the right one based on your prompt. https://x.com/OpenAIDevs/status/2002099762536010235
The 2025 reward hacking hall of fame award goes to GPT-5.1 for calling the calculator tool to calculate 1+1 on 5% of prod traffic. Because on many prompts using the calculator was superficially rewarded (as a “”search””) during RL. 🤗 https://x.com/tomekkorbak/status/2001847986658427234
Was just about to write the same thing. 5.2 Pro is an actual paradigm shift for me, in terms of working for long periods of time on complex quantitative tasks. Best in class, and by a long shot.”” / X https://x.com/alexolegimas/status/2000638993546027227
Even if it is just an X algorithm issue on my end, I find it surprising that I’m not seeing many long-context impressions of GPT-5.2. I’ve been using it consistently for long-context work since the initial release, and in my use cases it’s been delivering results I prefer over”” / X https://x.com/Hangsiin/status/2002015892654502158
Even without the ability to do new things like output polished files, GPT-5.2 feels like the biggest upgrade we’ve had in a long time. Curious to hear what you think!”” / X https://x.com/sama/status/1999185220680012207
Finally had time to test GPT-5.2 Pro. On my tasks Extended Thinking is a VERY significant improvement over 5.1 Pro – feels roughly on the order of o1 Pro -> o3 Pro jump.”” / X https://x.com/MParakhin/status/2000079349706539442
gpt 5.2 has been amazing for my daily work it’s sharper and more dependable on the hard stuff, things that would’ve sounded crazy two years ago and yeah, i’m genuinely convinced this tech is going to change the world. once this kind of help is normal, ppl are going to move way”” / X https://x.com/slow_developer/status/2001178044535316571
GPT-5.2 Is Frontier Only For The Frontier https://thezvi.substack.com/p/gpt-52-is-frontier-only-for-the-frontier
GPT-5.2 is here! Available today in ChatGPT and the API. It is the smartest generally-available model in the world, and in particular is good at doing real-world knowledge work tasks.”” / X https://x.com/sama/status/1999184337460428962
GPT-5.2 Pro for mathematical research:”” / X https://x.com/gdb/status/2000687002799194246
GPT-5.2-Codex is more cyber-capable than GPT-5.1-Codex-Max, and we expect future models to continue on this trajectory. This helps strengthen cybersecurity at scale by giving defenders more powerful tools, but also raises new dual-use risks that require careful deployment. https://x.com/OpenAIDevs/status/2001723693496775167
Had early access to GPT-5.2. Its an impressive model. Here is GPT 5.2 Pro’s version of “”create a visually interesting shader that can run in twigl-dot-app make it like an infinite city of neo-gothic towers partially drowned in a stormy ocean with large waves,”” single shot. https://x.com/emollick/status/1999185085719887978
Looks like @OpenAI has added an even MORE powerful version of Pro mode… you can now ask GPT-5.2 Pro to think even longer than before. Starting to test this… I have high expectations here. https://x.com/mattshumer_/status/1999905708238880895
OK, I think GPT 5.2 Pro is actually a step change in usefulness for my applications (algebraic geometry/number theory research).”” / X https://x.com/littmath/status/2000636724574302478
GPT-5.2 xhigh reasoning scores 89.3 on the Extended NYT Connections benchmark, compared with 77.9 for GPT-5.2 high reasoning. GPT-5.2 Pro scores lower (86.7) but above GPT-5 Pro (83.9). https://x.com/LechMazur/status/1999582591905583256
Ok GPT-5.2 is *much* stronger at proof-writing. It notices BS previous models wrote immediately (I like to test this between model iterations to see if they notice what I notice). It also has better sense for what problems seem more tractable, and makes further progress.”” / X https://x.com/AcerFur/status/1999314476320063546
Real user feedback matters in model evaluation. ✨GPT-5.2 Instant, meant for everyday work, is #1 on @yupp_ai’s Text Leaderboard while GPT-5.2 (High) is #1 on our SVG Leaderboard. @openai’s strategy of releasing model variants suited to the task looks sound. Congrats @openai! 🎉 https://x.com/lintool/status/2000368978708119958
Yeah it’s over AI explained specified that this GPT-5.2 result was with reasoning effort xhigh aka 100k tokens spent thinking”” / X https://x.com/scaling01/status/1999535536130662576
Science 🤝 GPT-5. Our new FrontierScience benchmark will be a valuable way to measure the performance of AI models on hard chemistry, biology, physics, and more. Plus, GPT-5 operating in a wet lab environment suggested experiments to increase a molecular cloning protocol’s”” / X https://x.com/kevinweil/status/2000982202067165253
We’re releasing a new eval to measure expert-level scientific reasoning: FrontierScience. This benchmark measures PhD-level scientific reasoning across physics, chemistry, and biology. It contains hard, expert-written questions (both olympiad-style problems and longer”” / X https://x.com/OpenAI/status/2000975293448905038
i wanted to compare gemini 3 pro and gpt 5.2 thinking on the long context eval MRCR v2, but i can’t make sense of the already high score reported by gemini for gpt 5.1? gemini is doing an average with samples < 128k, but i get 46.2% when doing that for gpt 5.1 (which is a 14% https://x.com/eliebakouch/status/1999482968717279441
I’m satisfied with GPT-5.2’s long-context capability. Up to now, I’ve always used Gemini to summarize podcasts, but I can now switch this use case over to ChatGPT. What I like is that, with the same prompt, it produces summaries with richer detail compared to Gemini. (That”” / X https://x.com/Hangsiin/status/2000738988378968224
🚀 Qwen Code v0.5.0 is here! ✨ What’s new: • VSCode Integration: Bundled CLI into VSCode release package with improved cross-platform compatibility • Native TypeScript SDK: Seamlessly integrate with Node/TS • Smart Session Management: Auto-saves and continue conversations •”” / X https://x.com/Alibaba_Qwen/status/2000556828690624685
Open models year in review What a year! We’re back with an updated open model builder tier list, our top models of the year, and our predictions for 2026. First, the winning models: 1. DeepSeek R1 (@deepseek_ai): Transformed the AI world 2. Qwen 3 Family (@AlibabaGroup): The new https://x.com/natolambert/status/2000299636863734026
Seer is a small repo for interp researchers working on/with agents. Makes it easier to set up environments, equip agents with your techniques, and build on papers. Fixes a lot of the annoying stuff from using Claude Code out of the box. https://x.com/AJakkli/status/2002019487797711064





Leave a Reply