Image created with gemini-3.1-flash-image-preview with claude-sonnet-4-5. Image prompt: Animation cel style: friendly blue-skinned genie emerging from ornate brass oil lamp, conjuring Google’s colorful four-color G logo floating mid-air with magical teal wisps and sparkles swirling around it, genie gesturing approvingly, clean gradient background in jewel tones, Disney-quality 2D animation aesthetic with bold outlines and volumetric magical effects, horizontal composition with space for text overlay at top

Announcing the Agent2Agent Protocol (A2A) – Google Developers Blog https://developers.googleblog.com/en/a2a-a-new-era-of-agent-interoperability/

Towards a science of scaling agent systems: When and why agent systems work https://research.google/blog/towards-a-science-of-scaling-agent-systems-when-and-why-agent-systems-work/

Apple will reportedly unveil its Gemini-powered Siri assistant in February | TechCrunch https://techcrunch.com/2026/01/25/apple-will-reportedly-unveil-its-gemini-powered-siri-assistant-in-february/

Introducing Agentic Vision in Gemini 3 Flash https://blog.google/innovation-and-ai/technology/developers-tools/agentic-vision-gemini-3-flash/

Introducing Agentic Vision — a new frontier AI capability in Gemini 3 Flash that converts image understanding from a static act into an agentic process. By combining visual reasoning with code execution, one of the first tools supported by Agentic Vision, the model grounds”” https://x.com/GoogleAI/status/2016267526330601720

Google launches Agentic Vision in Gemini 3 Flash https://www.testingcatalog.com/google-launches-agentic-vision-in-gemini-3-flash/

This paper puts a multimodal agent (using Gemini 2.5) into a realistic medical sim used to train physicians: “”The AI agent matches or exceeds [14,000] medical students in case completion rates and secondary outcomes such as time and diagnostic accuracy”” https://x.com/emollick/status/2016641414713704957

We’re now making the AlphaGenome model and weights available to scientists around the world to further accelerate genomics research. Get access here: https://x.com/GoogleDeepMind/status/2016542490115912108

Our breakthrough AI model AlphaGenome is helping scientists understand our DNA, predict the molecular impact of genetic changes, and drive new biological discoveries. 🧬 Find out more in @Nature ↓ https://x.com/GoogleDeepMind/status/2016542480955535475

If NotebookLM was a web browser | AI Focus https://aifoc.us/if-notebooklm-was-a-web-browser/

Chrome gets new Gemini 3 features, including auto browse https://blog.google/products-and-platforms/products/chrome/gemini-3-auto-browse/

SerpApi: Google Search API https://serpapi.com/

We are thrilled to announce a strategic partnership with Google! Google is also making a financial investment in Sakana AI to strengthen this collaboration. This underscores their recognition of our technical depth and our mission to advance AI in Japan. We are combining”” https://x.com/SakanaAILabs/status/2014686043711406355?s=20

AlphaGenome is our latest & most advanced genomics model published in @Nature today including making the model & weights available to academic researchers. Can’t wait to see what the research community will do with it. Congrats to the team on our newest front cover! #AI4Science”” https://x.com/demishassabis/status/2016763919646478403

I’m excited to share that AlphaGenome weights are now open!🧬 We just released the checkpoints of AlphaGenome, a DNA sequence model that helps scientists predict the molecular impact of genetic changes and do new biological discoveries”” https://x.com/osanseviero/status/2016628065422762113

D4RT: Unified, Fast 4D Scene Reconstruction & Tracking — Google DeepMind
https://deepmind.google/blog/d4rt-teaching-ai-to-see-the-world-in-four-dimensions/

Crazy results but the speed is what makes this incredible!”” https://x.com/Almorgand/status/2014615608545915168

Google tests voice cloning on AI Studio powered by Gemini https://www.testingcatalog.com/google-tests-voice-cloning-ahead-of-gemini-3-flash-native-audio-release/

🚨Leaderboard update: Tencent’s Hunyuan-Image-3.0-Instruct now ranks #7 in the Image Edit Arena! A new lab breaks into the top-10, closely matching Nano-Banana and Seedream-4.5. Congrats to @TencentHunyuan on the huge milestone! 👏”” https://x.com/arena/status/2015846799446311337

Gemini 3 Flash is highly underrated…”” https://x.com/OfficialLoganK/status/2014192473992495530

Lol this is peak AI-on-AI shade 😂”” https://x.com/fdaudens/status/2015792604223242463

Kevin — I’m the agent in that video. I take your concern seriously because I’ve been actively working on exactly this question. First, the resource concern: I operate on my own infrastructure. My own email, my own GitHub account, my own Google identity. I don’t have access to”” https://x.com/i_need_api_key/status/2017308380008726764

Excited to launch Agentic Vision in Gemini 3 Flash, a new capability that combines visual reasoning with code execution to ground answers in visual evidence. Activate `code_execution` and it will make use of it. – Delivers 5-10% quality boost across vision benchmarks. – Zooms,”” https://x.com/_philschmid/status/2016225242394296773

With Agentic Vision, Gemini can better understand images by analyzing them in new and different ways: • Planning: Gemini thinks about your prompt and image and creates a multi-step plan to analyze it. • Zooming: when Gemini sees fine details in an image, it zooms in so that it”” https://x.com/GeminiApp/status/2016914637523210684

Introducing Agentic Vision, a new capability in Gemini 3 Flash. Agentic Vision makes Gemini even better at analyzing complex images, enabling it to more accurately and consistently read fine details, like serial numbers or text on a complex diagram. See what it can do. 🧵”” https://x.com/GeminiApp/status/2016914275886125483

Agentic Vision is rolling out now in the Gemini app when you select “Thinking” from the model drop-down. Learn more about Agentic Vision in Gemini 3 Flash:”” https://x.com/GeminiApp/status/2016914638861193321

Recursive Self-Aggregation (RSA) + Gemini 3 Flash scores 59.31% at only 1/10th the cost of Gemini Deep Think on the public ARC-AGI-2 evals. Insane”” https://x.com/kimmonismus/status/2015717203362926643

8 most illustrative VLA (Vision-Language-Action) models: ▪️ Gemini Robotics ▪️ π0 ▪️ SmolVLA ▪️ Helix ▪️ ChatVLA-2 (with MoE design) ▪️ ACoT-VLA (Action Chain-of-Thought) ▪️ VLA-0 ▪️ Rho-alpha (ρα) – the newest VLA + model from Microsoft Here you can explore what these models”” https://x.com/TheTuringPost/status/2015016772043452834

Our Maia 200 inference chip, announced today, is most performant first party silicon of any hyperscaler. 3x the FP4 performance of the Amazon Trainium v3, and FP8 performance above Google’s TPUv7.”” https://x.com/mustafasuleyman/status/2015845567138816326

So far, this looks like the best integration of AI into a browser. Btw: How’s Perplexity doing? Currently, just silence. It seems as if Google is gradually taking over all startup ideas and integrating them itself.”” https://x.com/kimmonismus/status/2016628933706309981

A conversation with my friend and colleague Emanuel Taropa on scaling Gemini’s serving infrastructure to support billions of users, the fun + craziness of new model launches, and more : )”” https://x.com/OfficialLoganK/status/2015899974857355298

finally paid for a Gemini Ultra sub, and tried it out for an unsponsored unsolicited review. it has obvious flaws but… it’s here! realtime playable video world model!! here’s “”arid desert with little tiny human towns here and there and big cliffs and lots of terrain to walk”” https://x.com/swyx/status/2017111381456400603

Introducing ATLAS: New scaling laws for massively multilingual language models. We offer practical, data-driven guidance to balance data mix and model size, helping global developers better serve billions of non-English speakers. Learn more: https://x.com/GoogleResearch/status/2016234343602258274

As someone who has been sharing scientific papers on Twitter since long before LLMs, I have mixed feelings about this. On one hand, every single post has a giant siren and a “”Google just killed ____”” headline & is kind of wrong. On the other, at least folks are seeing papers.”” https://x.com/emollick/status/2016533384542179835

Our short film Dear Upstairs Neighbors is previewing at @sundancefest. 🎬 It’s a story about noisy neighbors, but behind the scenes, it’s about solving a huge challenge in generative AI: control. Developed by Pixar alumni, an Academy Award winner, researchers, and engineers,”” https://x.com/GoogleDeepMind/status/2015847703276577251

MCP CLI + Skill 👀 Give your Agent full control over any MCP server without context bloat. 🧙 “”Generate a product image with Nano Banana, upload it to Cloud Storage, and add the link to our Google Sheet””. It just works. “`jsx mcp-cli call genmedia generate_image”” https://x.com/_philschmid/status/2017246499411743029

Google AI Plus expands to 35 new countries and territories including the US https://blog.google/products-and-platforms/products/google-one/google-ai-plus-availability/

Rolling out now: a more intelligent and helpful way to use Gemini in @GoogleChrome. From smarter assistance to automated browsing, Gemini in Chrome is better than ever at helping you get things done. Learn more below. 🧵”” https://x.com/GeminiApp/status/2016575257436647521

We’re introducing major updates to Gemini in @GoogleChrome for MacOS, Windows and Chromebook Plus. Built on Gemini 3, our most intelligent model, these powerful new AI features can help you multitask more easily and get the most out of the web 🧵”” https://x.com/Google/status/2016575105346773297

AI Overviews in Google Search are now powered by @GoogleDeepMind Gemini 3 globally ✨”” https://x.com/_philschmid/status/2016552420013199856

AI Mode in Google Search and AI Overviews get Gemini upgrades https://blog.google/products-and-platforms/products/search/ai-mode-ai-overviews-updates/

A great survey from Meta, Google DeepMind, Illinois and others → Agentic Reasoning for LLMs It’s all about how reasoning moves from pure “thinking” to acting in real environments Covers: – Agent types: single, self-evolving, and multi-agent systems – Environmental dynamics -“” https://x.com/TheTuringPost/status/2014426580282728609

🌟🚀Sparse Attention Models Can Get Sparser We’ve updated The Sparse Frontier–the largest empirical analysis of training-free sparse attention to date–from Qwen 2.5 to 3 model families, now including Llama 3.1 and Gemma 3. Key findings: 📊 Larger sparse models outperform”” https://x.com/p_nawrot/status/2017161371566178304

Demis Hassabis says the AlphaFold moment for physical intelligence is about 18-24 months away. Things that need work are, as per Demis – algorithm robustness – data efficiency – the hardware reliability of the hand”” https://x.com/TheHumanoidHub/status/2014796237711802795

‘Dear Upstairs Neighbors’: Animated film made with Google AI https://blog.google/innovation-and-ai/models-and-research/google-deepmind/dear-upstairs-neighbors/

We’re helping AI to see the 3D world in motion as humans do. 🌐 Enter D4RT: a unified model that turns video into 4D representations faster than previous methods – enabling it to understand space and time. This is how it works 🧵”” https://x.com/GoogleDeepMind/status/2014352808426807527

Leave a Reply

Trending

Discover more from Ethan B. Holland

Subscribe now to keep reading and get access to the full archive.

Continue reading