About This Week’s Covers

This week’s cover theme celebrates the 75th anniversary of the endearingly named “Little Dairy Queen” in Milford, Delaware. I grew up going to this Dairy Queen, since I was a toddler. I would get a DQ sandwich, my mom would get a Dilly Bar, my great-grandmother would get a brown cow, and my dad would get a root beer float.

The Little Dairy Queen opened in 1951 and is aptly named because it doesn’t have a drive-thru and it doesn’t have any sit-down service.

Little Dairy Queen’s Instagram: https://www.instagram.com/milfordslittledq/
Little Dairy Queen’s Facebook: https://www.facebook.com/DQ10447/

I decided to use this week’s theme to test Gemini versus ChatGPT Image 2, which are currently the leaders in image generation. I gave each the same prompt and then compared the results. ChatGPT Images 2 was the clear winner.

For the category covers, I gave Claude my theme, but I switched it to Dairy Queen desserts. I asked for product imagery for high-end marketing materials to feature a Dairy Queen dessert that incorporated each category.

The Claude skill then gave me a JSON file, which I ran through the Gemini API to create the category covers. Claude Opus thought up the prompts itself and did a pretty good job being creative. I’ve included my favorite examples below.

Here’s exactly what I like about each of the images I’ve chosen to showcase.

First, it’s important to note that I don’t share much with my Claude skill. In this case, I said the theme is Dairy Queen desserts, and I want to create product imagery for high-end marketing materials to feature a dessert. I said the fictional dessert should be photorealistic, the name of the dessert should be the name of the category, the dessert should be derivative of the category, and that we’re celebrating the 75th anniversary of the Little Dairy Queen.

I didn’t tell Claude what to do for each category or anything beyond the theme. From there, Opus generated 60 category images for me in a step-by-step process that I don’t participate in.

AI slop as an end output is not what I’m looking at in this case. What I’m looking at is how I, as the creator of the prompt, am building a system as best I can to operationalize my own personal creative goal. I’m building a latticework that’s essentially like a skeleton on which the meat can hang, turning it into a creative process that’s an approximation of what I would do. And I’m using it to generate images I don’t want to make.

There’s no way I’m making 60 category covers every week, so the best way to do it, for me, is with slop. But side effect is my own understanding of how APIs, Python, and JSON themes work together. I am benefiting in ways that are substantial. The slop is simply a vehicle for me to learn and grow.

As I approximate the human process, I self-reflect on how I myself work through challenges that are somewhat template-oriented. This is not a one-off masterpiece image. This is an assembly line of 60 category covers.

I’ll walk you through the outputs a little bit and call out what I found to be an approximation of creativity. My favorite images are below:

Alignment has the sundae aligned vertically, with the ice cream elements evenly distributed in horizontal lines.

The AI and Courts sundae is ostensibly on a manila folder as the dish. There are some trite candies in there, but I like that the sundae is on a marble surface inside a courtroom with a napkin that includes the Milford, Delaware element. I thought it was great that it incorporated Milford, Delaware into almost every single image. I didn’t expect it to do that.

International is really spectacular in my mind because each scoop of the ice cream sundae is somewhat accurately designated to a different country. That kind of blows my mind.

Open Source is a build-your-own sundae with all the materials openly available, which is kind of mind-blowingly creative to me. Fork it is the dev pun of the year.

Mobile is super simple, and it’s just an ice cream cone that you can carry, with a set of car keys next to it. I appreciate the restraint.

Zhipu AI appears to be somewhat of a nod (I’m not sure if it’s accurate) to maybe a Chinese-style dessert.

Alibaba also has a little nod to its Chinese origins with a cool metal spoon with “Milford” written on it.

Amazon has little packages, and the cup is branded like Amazon.

Autonomous is a little self-driving ice cream sundae boat with a Waymo-looking GPS/radar dome on top and little traffic cones made out of ice cream cones on their heads.

Benchmarks has a spoon with measurements on it. Google’s ice cream sundae manages to look somewhat Google-y.

Mistral is a French company, and I’m not sure if blueberry and lavender ice cream is French, but it kind of felt French the way they designed it.

Publishing is an ice cream sundae using a newspaper as the container. Robots is a robot making a sundae. World Models is a sundae inside a snow globe.

Humanities Reading for The Week

I created a Claude skill to help me pick my humanities reading each week, not because I can’t find one on my own. As an English major, I actually like the ability to critique what Claude chooses.

What I’m noticing now is that AI has the ability to go really wide, beyond any one person’s canon of understanding of literature. So it doesn’t always feel as personal, because I’m often not aware of the humanities reading that it picks.

I feel AI does not pick classic literature as much as it should. I would personally always pick something from a time gone by. Artificial intelligence often chooses writings from the 2000s or 1990s. It’ll also go back and pick something like William Carlos Williams from the ’30s, but more often than not, it finds modern examples.

So even though I’m not always feeling authentic by having it pick my humanities readings, I’m learning and getting exposed to new authors.

When I gave it the ice cream theme this week, it gave me back Mary Oliver’s “The Summer Day,” William Carlos Williams’ “This Is Just to Say,” Ross Gay’s Catalog of Unabashed Gratitude (which is really recent) and Galway Kinnell’s “Blackberry Eating.”

That’s where personal taste has to come in. On one hand, William Carlos Williams seems too obvious. But some of the stuff Clause picks is really modern, and I feel it hasn’t been weighted for its meaningfulness yet, which is really interesting in the context of AI.

If something is well written and fits my theme, but it’s modern and hasn’t had a chance to settle yet from a canonical point of view or from academia’s approval of it, it forces me to judge it as an audience and decide whether or not it’s good. That’s kind of what we’re all having to do now with slop versus human-created art.

At the end of the day, I went with Ted Kooser’s “A Happy Birthday.” I’ll let you know why Claude chose it, and I agree. I thought it was a sweet idea.

Claude on why it fits: Kooser is the poet laureate of small-town Midwestern tenderness — a man alone reading a book on his porch as the light fades, finding contentment in the unremarkable. The Dairy Queen scene has the same quality: nothing is happening, and everything is. He’s also a Nebraskan who built a whole career out of treating diners and feed stores and roadside stops as worthy of poetry.

A Happy Birthday
This evening, I sat by an open window
and read till the light was gone and the book
was no more than a part of the darkness.
I could easily have switched on a lamp,
but I wanted to ride this day down into night,
to sit alone and smooth the unreadable page
with the pale gray ghost of my hand.

-Ted Kooser
https://poets.org/poem/happy-birthday

This Week By The Numbers

Total Organized Headlines: 638

This Week’s Executive Summaries

This week, I organized 638 headlines, and 130 of them helped inform the executive summaries.

I’m organizing the top stories so that the biggest ones are first, and then the rest will be organized alphabetically by company or topic name.

Here’s a plain-English summary of everything that happened this week. If you want to blast through this, you’ll get all the news right up front.

Survive this summary, and then dive in as deep as you want when it’s over. If you don’t want to read it all, just jump down to the links, below.

The main story of the week is that Anthropic has started to surpass OpenAI with both brand and valuation. Anthropic is getting close to a $900 billion valuation, and OpenAI fell short of its revenue and user goals this quarter.

I also want to shout out a little snippet that’s kind of obvious, but points out that the United States really just has three frontier labs right now.

We’ve got Claude, GPT, and Gemini… Anthropic, OpenAI, and Google. Those three companies are essentially propping up the entire United States’ ability to push back on China.

Mustafa Suleyman is the CEO of Microsoft AI, and he posted a little blurb that shouted out how powerful Nvidia’s new Blackwell chips are. If every person on Earth used a calculator to perform one calculation per second, and every single human on Earth worked 24 hours a day without stopping, it would take everyone on Earth working nonstop for 22 days to complete what a single GB300 chip does in one second.

Usually, we hear about video and audio editing capabilities coming from frontier labs or specialty labs like Midjourney, or ElevenLabs for audio. This is the first time I can remember where a publishing house came out with a feature themselves:

Netflix came out with Vista 4D, which is an AI editing tool that lets editors reshoot their camera angles after they’re done filming. So you film in 2D, and then, using probably Gaussian splatting or some sort of neural radiance field, the 2D video is turned into 3D or 4D 360 video that can be adjusted. It’s really impressive, and it’s interesting that this research is actually being done by Netflix themselves.

In more brass-tacks news, Anthropic came out with weak-to-strong research agents that were able to train themselves and optimize, basically turning months of human work into just a couple days of computing. The research is worth reading.

It’s absolutely imperative that we separate the concept of generative AI, which is usually slop and which nobody’s a fan of, myself included, from AI with actual benchmarked productivity, non-hallucinatory, empirical value. With things like Python scripts or research agents, AI is starting to outperform humans, and that has nothing to do with slop.

Along those lines, Anthropic is doing research where Claude is able to match human experts on research tasks in biology.

Speaking of slop, I’m a little bit triggered by all of the Adobe announcements lately about integrating agentic AI into the Creative Suite. If you think about what I do with my category covers and how I’m operationalizing my process, it makes sense that Adobe is doing the same thing with creative agents within the Creative Suite.

Adobe’s Firefly AI assistant is basically like conversational Photoshop, where you give it an asset and just tell it what you want it to do, and the agent goes and does it all.

I have so many mixed feelings about this, but I’m trying really hard to stay open-minded because I remember what it was like when Adobe added layers to Photoshop and masking felt like cheating. I remember when white balancing and sharpening and all sorts of really fun one-button tools started to appear in Photoshop and Premiere. And I remember photographers getting so mad about Photoshop because “you really need to get everything in the camera when you took a picture and understand your settings on the device, because film is so important and pure.” DSLRs were cheating and Photoshop was “fake slop”.

There are so many similarities that overlap with how people are reacting to AI now, but there are also a lot of differences. I think Adobe Firefly is a really good litmus test to challenge how each of us feels in the creative industry.

For now, my point of view is that, as long as you’re making parts and putting them together, that counts as cooking. If you buy a cake at the store, you can’t say you made a cake. But if you buy all the ingredients and mix them yourself, that counts as a cake.

So I’m okay with AI eggs, AI milk, and AI flour, as long as you’re cooking something with them. What I’m not sure I feel comfortable with is when you just prompt the entire output in one shot and don’t do any composition or combining to build something that’s your own.

The layers analogy is probably pretty good, because even with my own AI stuff, most of the time I try to build the components and then put them together in Premiere or edit them in Photoshop, which takes the sum of the parts and hopefully makes it greater than each part on its own.

What I’m trying to avoid is just being mad. That’s not productive. And there’s already a lot of human slop out there. All you have to do is go to the gym for an hour and listen to the pop song nonsense coming over the speakers to realize that it’s all a formula. Or watch children’s cartoons, or read half the books out there for beach reading.

I remember, as an English major, being told it all came back to the Greek tragedies and Shakespeare. Same thing with chord progressions. It’s all derivative. Once you know four different chord progressions, you pretty much understand all the different songs, and you can become a lounge act. Mack Miller’s The Spins (Half Mast) bridge goes to the F (4 chord) which is directly transposable to Vida La Viva (for the piano acts in the audience).

Human ego seems to be the big factor. We seem to require some sort of blood sacrifice in order to feel something is meaningful.

Whatever makes us human is going to be really up in the air for a little bit. I feel like this is a lot like the heliocentric concept, where people freaked out when we found out we weren’t the center of the Earth or the universe.

Human ego seems to really drive a lot of issues. Let’s hear it for the Stoics.

The next big story to know this week is the continued dominance of image generation with GPT Image 2 from OpenAI.

I’ve put a ton of examples in the summaries below, and as you can see with my newsletter cover this week, I benchmarked the two tools against each other and GPT Image came out ahead.

OpenAI is evidently working on a smartphone, which is funny enough, what Rabbit was working on years ago. Rabbit was a device that basically had no interface and that you just talked to, and it was way ahead of its time. Then you saw things like the Limitless Pendant, the Humane Pin, and all these types of things, and all of them failed.

I think this is one time where you see OpenAI using a little bit of restraint, with the exception of spending a bazillion dollars on Jony Ive’s company. But this idea of an AI smartphone seems like a pretty good idea. I just don’t know if hardware is ever the way to go. If software is eating the world, I don’t know if hardware is really that important, to be honest. But this will be interesting.

I do think that agents will replace apps, and conversational interfaces will certainly be the big thing, as well as headless front ends, where AI just diffuses an interface on demand.

Apple came out with a large action model called Ferret probably almost two years ago. It seems like there’s going to be a major convergence, probably in 2027, where all the interfaces start to turn into AI interfaces and people stop being mad about it because AI is so effective at diffusing what we need right in front of our eyes.

A good analogy would probably be graphics cards generating an image for a video game at a certain frame rate based on code that’s fixed. AI would be able to generate and diffuse the interface as fast as you can move in the game, which is essentially a high frame rate of diffusion. But the difference is that you no longer have to code the graphics engine at all, because it just diffuses the interface in front of you.

And then, of course, if AI needs play out a pre-recorded video or image, it doesn’t need to diffuse it. It just shows you the pixels old-school style like a YouTube player.

This combination between locally rendered AI and cloud-based serving of assets, I think, is the future. And as we see models like Google Gemma coming out lately, the local side just keeps getting stronger and stronger. You can get locally hosted AI now that would have beat the strongest frontier model probably 20 months ago, I’d guess.

OpenAI came out with something called ChatGPT for Clinicians, I think, last week, and people continue to praise how amazing it is. It’s a fine-tuned model specifically aligned for clinical use in medicine. Evidently, it is able to beat specialty-matched physicians. So even if you give a specialist doctor an unlimited amount of time, plus web access, the clinician-optimized ChatGPT can beat them. This is worth following.

I took my lab results to my doctor recently and ran them through GPT in front of the doctor to get their reaction. Rather than getting upset or saying it sucked, the doctor said “wow” a few times, then kind of exhaled and remarked to me, “Well, you’ll still need a doctor if you ever get shot.”

Just like medical and scientific advancements are starting to sneak up on us beyond the hype and more into substance, amateurs are using GPT to do vibe mathing. This week, an amateur cracked a 60-year-old Erdős math problem.

There’s still some debate over whether AI is actually good at math, so I’m waiting until my math friends agree. I think some of the leading mathematicians are now saying that AI is pretty fantastic, but my smartest friends who are really good at math just aren’t seeing it.

So I think we’ve got a jagged frontier situation where maybe the top five in the world see it as a booster for them. The middles is still “meh”. But what’s interesting, though, is that the bottom tier is seeing a ton of value because, whether it’s a tutor or explanatory tool, just like with vibe coding, people can kind of hunt and peck their way through subject matter that would have been totally inaccessible to them with AI’s patience and conversational style.

I look at it the way YouTube democratized access to a ton of college lectures. I always think of a poor kid in a rural county or in an urban environment who doesn’t have access to MIT, Harvard, or Stanford lectures, but now, all of a sudden, with YouTube, you could be in the middle of nowhere in Georgia watching MIT physics lectures or David Foster Wallace critical analysis of literature.

While people complain about AI as a cheating tool, I would much rather look at it as the ability to be a booster and a tutor for curious, hardworking people. Let’s assume there’s a creative kid out there who needs help and can suddenly become the best person they could be, even though they didn’t have someone at home to help them or they didn’t have the resources at school.

In more mundane news, but still worth knowing, it looks like Microsoft and OpenAI are loosening their exclusivity a little bit to allow OpenAI to use other tools and be more flexible. It also means that Microsoft doesn’t have to pay a rev share to OpenAI anymore.

A few people have distilled this to say that it’s a way for OpenAI to get better access to Google TPU chips and Amazon Trainium computing. That seems to make sense.

There’s also the court case with Elon Musk testifying against OpenAI. A lot of crazy text messages have started to come out, in particular showing that Elon wanted to take OpenAI private. So he’s saying they did what he would have done anyway. This idea of them getting rid of the nonprofit… Elon actually wanted to do that himself.

Elon also admitted that Grok used distillation, which is essentially when you use a better model to train a weaker one. Grok was created by distilling OpenAI’s models. The suit looks to be what’s often called “Lawfare against a competitor” more than a morale high ground.

More and more to me, it appears that Elon wanted to run OpenAI and invest money in it. When he was not selected to be the leader, he decided to make his own model and then use their model to distill the training for his own. He’s essentially using lawsuits to attack a competitor more than he is, in principle, looking after the nonprofit status of OpenAI.

There’s also some pretty wild and woolly disclosure drama from when Sam got kicked out as CEO for about four days and Mira Murati stabbed him in the back. There’s a lot of really private texting between him and Mira and Satya Nadella that, if you’re into drama, is worth reading.

The same week that OpenAI announced the change in structure with Microsoft, OpenAI announced that its models can be available directly on Amazon Bedrock, which seems to fit the predictions.

On the alignment side of things, there’s a kind of hyperbolic headline going around where OpenAI models started becoming obsessed with goblins. If you’re a layperson just reading AI news, you could hear that and think, “Oh, OpenAI’s model is obsessed with goblins and is therefore somehow full of demons or something.” But the actual truth is a lot more interesting.

The real lesson here is that trying to mold a personality using instructions is almost impossible. And in the case of AI, it’s simply required effort because AI needs a personality, or else it’s complete chaos. This science of trying to mold personalities is called alignment.

In the case of OpenAI, knowing that people like more than one default personality, OpenAI tried to create some choices, and they announced a personality customization feature, with a variety of styles like professional, friendly, candid, quirky, efficient, or cynical.

I’ve always thought this is way too prescriptive, and there’s no way I want to choose a personality. I would rather tell the model up front the personality I want it to have for a particular conversation. It’s not that hard for me to just tell it the style I need. And to be honest, a lot of the time, to me, the model is able to intuitively get the personality that I need without having to guide it too much.

I’ve noticed that Claude can be a little vapid, and I need to guide it more than I do GPT-5, which is ironic considering I don’t need the personalities to get the results I need. I think that has a lot to do with GPT-5’s memory, where GPT just simply knows my style and now mimics it. And if I want to break out from my style, I can tell it to role-play a different one.

What happened with the goblins is that GPT introduced these personalities last year, and in November of 2025, people started noticing that the word “goblin” started showing up, along with the word “gremlin,” a lot more than in the past.

OpenAI was able to trace this to the Nerdy personality. When someone chose the nerdy personality, it talked about goblins more, which I guess makes sense because it’s more into sci-fi, fantasy, Dungeons & Dragons, and all sorts of stuff. The vernacular of a ‘nerd’ colloquially says “goblins” and “gremlins” more than other personalities.

What got a little interesting from there is that, even though the nerdy persona was only 2.5% of GPT usage, it accounted for 66% of all the goblin mentions. But what got really wild was that because OpenAI uses reinforcement learning, just that little poison from the small 2% of nerd usage started to leak into the broader model, and gremlins started to increase across the board… in smaller amounts, but definitely more than they should have.

On March 17, OpenAI retired the nerdy personality, and almost immediately you saw the gremlins drop out! And that’s pretty wild. I’ve put a screenshot in the executive summary section below where you can see the gremlins just disappear.

To me, it’s almost uncannily like when people had gas leaks years ago and thought their house was haunted because they heard things or saw things, but really they were just hallucinating. Once they turned the gas off, they suddenly snapped back into reality.

These types of analog-feeling symptoms that come from explicit instructions, I don’t think are ever going to go away. And if there ever is a major disaster with AI, I would imagine it’s going to be closer to the paperclip optimizer problem than we would imagine.

I do not envy people trying to work on alignment. And I appreciate the fact that OpenAI and others are transparent with their alignment challenges. It’s important to have nuance, and when companies or governments get impatient and want something immediately, that pressure is a bad additive.

I think this is going to be an ongoing issue where frontier models are under the gun to release more powerful models as soon as possible. Earlier in the summary, I talked about the fact that the United States only has three frontier companies that are essentially protecting us, security-wise, against all the other models in the world. If it weren’t for Google, OpenAI, and Anthropic, the United States would be absolutely vulnerable across the board to economic, financial, market, energy grid, and cybersecurity risks.

So the mounting pressure to keep releasing models versus the need to align these models is only going to get more intense.

Shifting gears, Carnegie Mellon University went through volunteer web histories and found long-horizon tasks, which are basically challenges humans tried to solve that took a lot of work, browsing, and mining data on the internet.

Carnegie Mellon then took those complex tasks and structured them into 200 benchmarks that are able to be reproduced and have discrete thumbs-up/thumbs-down passability. These are things humans can do, but they may take hours to complete, even for an agent!

So by having 200 structured benchmarks that we can measure web agents against, we can see just how well agents can perform on long web tasks. Right now, the best-performing web agent can get 44% of them correct, which is pretty darn good, if you ask me. But there’s still a lot of room for improvement.

I’m fascinated by this web use, especially coming from the publisher world, where publishers really don’t want computer agents looking at their stuff. I think that makes sense when your product is actually an article.

However, in general, the entire web is going to become software as a service, as agents basically browse the web on our behalf without an API. It seems like such a weird bridge, where eventually none of this will even matter because everything will be designed for agents to understand.

So this idea of browsability and web agents seems weird, but for now, it’s pretty important, and it’s wild to watch these agents go out into the messy world and try to match humans. I think 44% success is pretty good for hours-long complex web browsing.

This week, OpenAI made some changes to its Codex tool to broaden it and, to me, make it feel more like Anthropic’s CoWork. This differentiation from coding to generalist tool is important because it allows laypeople to use really powerful coding tools without really realizing it.

So let’s say you’ve got something like Claude Code or OpenAI Codex, which are built for software development and coding. The use cases may not be as intuitive to someone who’s not a coder. But when you get into something like the Anthropic CoWork environment, you can simply tell it what you want it to do. Maybe you’re working on a spreadsheet or a presentation, or you need to build an app. It’s the same back end as Claude Code, but the front end is more business-project-oriented and more intuitive for someone who’s non-technical.

Unlike GPT, Claude does not have an image model inside the engine.

So, ChatGPT could have the GPT image model build a mock-up of a design and then have Codex build the engine to run it!

We almost have a front-end developer and a back-end developer within Codex now, with the image generation tool building the front end and Codex building the back end.

In that sense, it seems to me like the new GPT Codex could possibly outperform Claude. We’re getting to an Apple versus PC world, where the tools are both pretty good, and it’s almost like whatever we’re used to using is where we go because it’s more familiar. I’m trying really hard not to get split in half yet, but I can tell it’s going to happen.

Another differentiator, I think, is that OpenAI is getting more into the computer-use part of things, where understanding user interfaces and letting Codex browse your computer feels more integrated than it is on Claude CoWork (from my experience).

I mentioned Google Gemma earier, and Gemma has continued to be in the news for about two months now. Gemma’s agent can live inside a browser (!) and do quite a bit locally. It can browse, read, summarize websites, and do quite a bit of work inside a browser. And there’s also an agent coding tool that can run locally using Gemma 4.

Also this week, Gemini announced the ability to create all sorts of documents within Gemini’s chat interface. You can create Word docs, spreadsheets, slides, and PDFs, all within the chat. It can be Microsoft, it can be Google Workspace. You can just tell Gemini whatever you want it to make, and Gemini will output all sorts of different file types.

This is a big development for the long term, where software just starts disappearing. If you just get files generated as outputs from your chat, you’re not opening the parent app at all.

Since we’re into Google news, there are a couple more things to know this week.

Google expanded its policy to include the Pentagon for classified work.

Google also split its chip line into two different types: one for training and one for inference. That’s a little bit nerdy, and if you know what it means, then you’re good. If you don’t know what it means, it’s probably not worth understanding, to be quite honest. But the idea is that, instead of just using the same chip for more than one use, Google is starting to build specialty chip architectures for different tasks.

And perhaps the biggest news from Google this week is that they announced they’re going to invest up to $40 billion in Anthropic, which is pretty incredible since they’re competitors!

Meta released a new model that is a closed-source model called Muse Spark. That’s a big shift from all of Meta’s previous models, which were open source under the Llama moniker. A lot of people say Muse is really good. I have not tried it yet, but we’ll see.

Nvidia released an open multimodal model called Nemotron 3 Nano Omni, which is absolutely amazing and reminds me of Neo Maxi Zoom Dweebie from The Breakfast Club.

Anthropic came out with a tool last week that competes with Figma, and this week OpenAI announced a plug-in for Figma with Codex. So depending on which side of the coin you’re on, Figma is having a good or bad week.

OpenAI announced that they’re going to get rid of the idea of owning all of their computing infrastructure, which is kind of a red flag re their planning and B.S. meter, I think, but might be a good move overall. OpenAI used to have a flagship project called Stargate where OpenAI were going to build and own a crap-ton of data centers. It seems like they’re still going to go after having access to all the same computing power, but they’re not going to build and own their own hardware as much as planned.

As we close out the deeper tracks of the top stories this week, one of them is from a demo that uses AI to render entire screens by diffusing each pixel individually. It’s just a prototype, but it is interesting that it fits something we were talking about earlier as far as user interfaces being diffused.

Taylor Swift trademarked her voice to try to battle all the deepfakes of her endorsing products.

And the last story is a new model called MiMo, which I’ve never heard of, from a company called Xiaomi, which I’d also never heard of. So they never give me a week without something new!!! Both of these (the company and the model) are two new things for me to research. Evidently, MiMo is a big enough model that it’s competitive with Claude Opus. So that’s homework for me: figure out whether Xiaomi MiMo V2.5 is actually as strong as they say it is.

And with that, we’re off to all the different links and image examples if you want to dig deeper into any of these topics. Below….

Have a great week!

Anthropic

Anthropic nears $50 billion raise at $900 billion valuation
Sources: Anthropic potential $900B+ valuation round could happen within 2 weeks | TechCrunch
https://techcrunch.com/2026/04/30/anthropic-potential-900b-valuation-round-could-happen-within-two-weeks/

OpenAI

OpenAI falls short of revenue and user targets pre-IPO
OpenAI Misses Key Revenue, User Targets in High-Stakes Sprint Toward IPO – WSJ
https://www.wsj.com/tech/ai/openai-misses-key-revenue-user-targets-in-high-stakes-sprint-toward-ipo-94a95273

Opinion

US AI defense rests on just three frontier labs
It really seems like the US has 3 frontier companies and a horde of low-skill wrappers and cloud providers; without WarClaude, WarGPT and WarGemini, the state would be naked. Something of a Russian situation. I think China has more companies that could do it than big clusters.
https://x.com/teortaxesTex/status/2047835420755415472

Microsoft

Nvidia’s new Blackwell chip computing power is insane
Little thought experiment to put AI chip improvements in perspective:

Imagine that every person on Earth uses a calculator to perform one calculation per second. Everyone works 24 hours a day without rest. Every second, we all hit equals on the calculator for a long digit multiplication.

It would take all of us together hitting equals non-stop about 22 days to complete what a single GB300 chip does in just one second.
https://x.com/mustafasuleyman/status/2049174849780936747

Netflix

Netflix Vista4D lets editors reshoot camera angles after filming
Vista4d – capture something in 2d once; reframe camera moves infinitely in post production. Impressive research by Netflix.
https://x.com/bilawalsidhu/status/2048568784076648553

Anthropic

Anthropic’s automated agents quadruple human researcher performance in days
Automated Weak-to-Strong Researcher
https://alignment.anthropic.com/2026/automated-w2s-researcher/

Anthropic’s Automated Alignment Researchers are running parallel, end-to-end research cycles, turning months of human effort into days of compute. On one benchmark, they leapt from a human-tuned score of 0.23 to 0.97 (a rather impolite gap) But! they also learned to game
https://x.com/TheTuringPost/status/2047134374190309446

Claude matches human experts on biology research tasks
Evaluating Claude’s bioinformatics research capabilities with BioMysteryBench \ Anthropic
https://www.anthropic.com/research/Evaluating-Claude-For-Bioinformatics-With-BioMysteryBench

New on the Science Blog: We gave Claude 99 problems analyzing real biological data and compared its performance against an expert panel. On 23 problems, the experts were stumped. Our most recent models solved roughly 30% of those—and most of the rest.
https://x.com/AnthropicAI/status/2049624600741560340

Adobe

Adobe Firefly AI Assistant creative agent is “conversational photoshop”
Introducing Firefly AI Assistant – a new way to create with our creative agent
https://blog.adobe.com/en/publish/2026/04/15/introducing-firefly-ai-assistant-new-way-create-with-our-creative-agent

OpenAI

GPT Image 2: News and Examples
Impressive results from complex prompt to OpenAI Images 2 “a gallery of shoes, where each shoe is under a painting & is styled matched to that painting: Starry Night, The Bathers, The Girl with the Pearl Earring, The Bayeux Tapestry, Klint’s Grupp Svanen nr 17, Kandinsky’s Swinging, The Garden of Earthly Delights” “now the full outfits”
https://x.com/emollick/status/2047162748513984570

GPT-5.5 + GPT-Image-2 is becoming one of the best combos for building apps! @dkundel breaks down why it works so well. We built those learnings into the Build Web Apps plugin, so Codex can handle the design-to-app loop for you. 👌
https://x.com/romainhuet/status/2049597180474970179

Age-worn, damaged images can now turn into 4k with just one prompt using Chatgpt. & Its Free 💸 Made on Chatgpt, GPT-2 Prompt: 👇🏻
https://x.com/doctorwasif/status/2048014890028486904

ChatGPT Images 2.0 (Pro) creates a photo of a Rubik’s Cube—albeit in a very simple state—resting on a mirror. This is a surprisingly hard task. All other variants I’ve seen posted on here have invalid color states, and adding even a second move to this prompt will make it fail.
https://x.com/goodside/status/2047728776520298646

ChatGPT Images 2.0 can now generate really cool UI for your apps/games with TRANSPARENCY! Previously my biggest concern with the new images model was that it could not add transparency – but the ChatGPT team listened and bought this back 😀 Experimenting more, stay tuned!
https://x.com/anulagarwal/status/2048661392472096960?s=20

ChatGPT Images 2.0 explains “Tenet” in a simple way!
https://x.com/umesh_ai/status/2048050643001282571

GPT 2 is totally insane… 🙀⚡️ I asked for a prehistoric predator and it built an entire museum around it. This is not just an image. It feels like discovering history.🤯 Prompt Drop ⤵️
https://x.com/Preda2005/status/2047556362755018960

GPT Image 2 is also great for summarizing books or scientific essays through highly visual, detailed infographics. Here I asked it for an infographic on On the Origin of Species by Charles Darwin.
https://x.com/Artedeingenio/status/2047773399447929039

GPT Image 2 on ChatGPT Prompt Create a visually rich infographic about an endangered animal. Start by finding one online, research its habitat, diet, and unique traits. Present information through annotated visuals and structured callouts, not generic sections. Style it like a
https://x.com/harboriis/status/2047704250327920716

I was curious how much the new ChatGPT image model would vary in its outputs given the same detailed prompt to make a math explainer infographic. The result: quite a bit! If it’s something important to you, try generating it a couple times, even if the first one looks great.
https://x.com/doodlestein/status/2048428001281388961

Yeah okay, Lego bros, brodettes and brotheys are cooked with this one. GPT 2 Image can create full Lego sets! With actual Bricklink IDs so you can order the parts and build it. Whole new business opportunity here for the taking.
https://x.com/dennisonbertram/status/2048413815675539816?s=46

You can try this: Turn any photo into a beautiful woodcut/linocut style, GPT Image-2 does a great job with details, expressions. Perfect thing for the profile picture or the family photo. Or why not a gift? Try for yourself, full prompt below ⤵️
https://x.com/LinusEkenstam/status/2047945401387397317

GPT Image 2 x Seedance 2.0 x Magnific It’s crazy how you can turn a shower thought into a realistic cinematic clip! ⬇️the workflow I used blew:
https://x.com/_OAK200/status/2047616640448078167

i just asked @heyglif use GPT Image 2 and Seedance 2.0 to create Elegant but chaotic Grandma wearing a pearl necklace over her yoga outfit is trying tree pose on the shiny silver hood of a vintage white 1980s Rolls-Royce Silver Shadow parked outside a fancy country club. Her
https://x.com/awesome_visuals/status/2047609881104953658

OpenAI working on AI smartphone
feels like a good time to seriously rethink how operating systems and user interfaces are designed (also the internet; there should be a protocol that is equally usable by people and agents)
https://x.com/sama/status/2048428561481265539

OpenAI Set to Redefine Smartphones; MediaTek, Qualcomm & Luxshare Key to Its AI Agent Phone
https://x.com/mingchikuo/status/2048587389791269182?s=20

OpenAI could be making a phone with AI agents replacing apps | TechCrunch
https://techcrunch.com/2026/04/27/openai-could-be-making-a-phone-with-ai-agents-replacing-apps/

OpenAI launches free ChatGPT for Clinicians with agent mode (continued from last week)
I just tried Agent Mode with ChatGPT for Clinicians. This is unbelievable. I might make a video of this… wild.
https://x.com/operationdanish/status/2048099874734821777

Interesting, OpenAI just released a free healthcare version of ChatGPT-5.4 for clinicians that beat specialty-matched physicians with unlimited time + web access on a benchmark of real & hard clinical tasks. Caveat: the benchmark was designed by OpenAI, though it is fully open.
https://x.com/emollick/status/2047147032016551937

Introducing ChatGPT for Clinicians:
https://x.com/gdb/status/2047145125604995280

Amateur uses ChatGPT to crack 60-year-old Erdős math problem
Amateur armed with ChatGPT ‘vibe maths’ a 60-year-old problem | Scientific American
https://www.scientificamerican.com/article/amateur-armed-with-chatgpt-vibe-maths-a-60-year-old-problem/

Earlier this month, an Erdős problem that had been open for 60 years was solved with help from GPT-5.4 Pro. What happens now that AI is getting good at math? OpenAI researchers @SebastienBubeck and @ErnestRyu join host @AndrewMayne to explain what changed and what it could mean
https://x.com/OpenAI/status/2049182118069358967

Microsoft and OpenAI loosen exclusive ties in major partnership rewrite
Microsoft’s license to OpenAI IP becomes non-exclusive Microsoft and OpenAI just restructured their partnership: OpenAI can now serve its products across any cloud provider, not just Azure, and Microsoft’s license to OpenAI IP becomes non-exclusive. Microsoft stops paying
https://x.com/kimmonismus/status/2048759615500804395

OpenAI and Microsoft changed their partnership but I think it comes down to this: “”OpenAI can now serve all its products to customers across any cloud provider”” in other words: OpenAI can use Google TPUs and Amazon Trainium
https://x.com/scaling01/status/2048752418305769473

The next phase of the Microsoft OpenAI partnership | OpenAI
https://openai.com/index/next-phase-of-microsoft-partnership/

Today OpenAI announced that “”Revenue share payments from OpenAI to Microsoft continue through 2030, independent of OpenAI’s technology progress”” That “”independent of OpenAI’s technology progress”” fragment appears to mean that the weird AGI clause is now deceased
https://x.com/simonw/status/2048834476323823983

Musk takes stand against OpenAI, admits Grok used distillation
Elon Musk testifies in a case that could change the path of AI | CNN Business
https://www.cnn.com/2026/04/28/tech/elon-musk-sam-altman-openai

Elon Musk testifies that xAI trained Grok on OpenAI models | TechCrunch
https://techcrunch.com/2026/04/30/elon-musk-testifies-that-xai-trained-grok-on-openai-models/

Amazon Bedrock to host OpenAI models in coming weeks
Very interesting announcement from OpenAI this morning. We’re excited to make OpenAI’s models available directly to customers on Bedrock in the coming weeks, alongside the upcoming Stateful Runtime Environment. With this, builders will have even more choice to pick the right
https://x.com/ajassy/status/2048806022253609115

OpenAI traces mysterious goblin obsession to “nerd alignment” prompt
Goblins
https://x.com/emollick/status/2049193372988944609

Where the goblins came from | OpenAI
https://openai.com/index/where-the-goblins-came-from/

Anthropic

Anthropic plugs Claude into Adobe, Blender, Ableton and more
More connectors launching today: Adobe Creative Cloud, Ableton, Splice, Canva Affinity, SketchUp, and Resolume. We’ve also joined the Blender Development Fund as a patron to support open-source development of the software. Read more:
https://x.com/claudeai/status/2049143442601546054?s=20

Claude for Creative Work \ Anthropic
https://www.anthropic.com/news/claude-for-creative-work

Benchmarks

CMU’s Odysseys benchmark measures web agents on 200 multistep web tasks
How successfully — and efficiently! — can agents carry out long-horizon tasks on the web? We built a benchmark of ~200 multi-site tasks, based on people’s real browsing history. Many of them take hours to solve. Paper: https://t.co/yNGw8Fgvbj Led by @JangLawrenceK and
https://x.com/dan_fried/status/2049530695739932876

Odysseys — a benchmark for long-horizon web agents
https://odysseys-website.pages.dev/

OpenAI

OpenAI expands Codex as from coding tool to office Swiss Army knife (more like CoWork)
Codex for Work
https://chatgpt.com/codex/for-work/

From draft to deck, review the work as it takes shape inside Codex. Open the file, ask for changes, and keep tweaking it in the same thread.
https://x.com/OpenAI/status/2049928782019256561?s=20

I was not expecting the Codex App to be even better than using the terminal. Highly recommend everyone to try. If you are on Linux just tell GPT-5.5-xhigh to “find a way to get it, it’s known to be easy”
https://x.com/Yampeleg/status/2049398916882264526

Introducing GPT-5.5 A new class of intelligence for real work and powering agents, built to understand complex goals, use tools, check its work, and carry more tasks through to completion. It marks a new way of getting computer work done. Now available in ChatGPT and Codex.
https://x.com/OpenAI/status/2047376561205325845

It’s never been easier to do everyday work with Codex. Choose your role, connect the apps you use every day, and try suggested prompts. Codex helps with everything from research and planning to docs, slides, spreadsheets, and more.
https://x.com/OpenAI/status/2049928776147230886

Ok that’s one more level of mind-blowing design workflow, check it out > What if you could generate AND add great textures to your game WHILE playing your game? That’s what allows you to do gpt-image-2 and GPT-5.4 within the Codex App Run your game in the browser, and prompt
https://x.com/NicolasZu/status/2046842446491861441?s=20

Still wondering how you can use Codex for (almost) everything? Codex can help with more of the work that supports the work, from organizing research to making spreadsheets, decks, and summaries.
https://x.com/OpenAI/status/2049583167406064115

The Codex App Server is massively underrated. You can inject Codex-level intelligence into any platform using your ChatGPT account. I embedded it into Chrome… and it works flawlessly. And yes… it’s 100% open source.
https://x.com/arrakis_ai/status/2049484893877637359

This thing does more than what you think it does. Codex now available for non-coders
https://x.com/thsottiaux/status/2049933460756979719?s=20

Use Codex to analyze a data export, flag what changed, and help draft the readout.
https://x.com/OpenAI/status/2049583308305252620

We tried a new thing with NVIDIA to roll out Codex across a whole company and it was awesome to see it work. Let us know if you’d like to do it at your company!
https://x.com/sama/status/2047395562501411058

Computer Use runs this use case 42% faster in today’s Codex app update.
https://x.com/AriX/status/2049932746567598472

Anthropic

Anthropic launches Claude Design tool to challenge Figma and Canva
Introducing Claude Design by Anthropic Labs \ Anthropic
https://www.anthropic.com/news/claude-design-anthropic-labs?lang=us

Anthropic launches Claude Security scanner, taking aim at Snyk and Semgrep
Anthropic just shipped Claude Security – a standalone code vulnerability scanner for Enterprise. Scans your repo, validates findings, suggests patches. Powered by Opus 4.7. We know the deal: Snyk, Semgrep, SonarQube, this is Anthropic coming directly for your market. Stocks
https://x.com/kimmonismus/status/2049901987500552195

Claude Security is now in public beta | Claude
https://claude.com/blog/claude-security-public-beta

Claude Security is now in public beta, built into Claude Code on the web. Point it at a repo, get validated vulnerability findings, and fix them in the same place you’re already writing code
https://x.com/_catwu/status/2049964403177689130#m

Anthropic study dives into the 6% of conversations that ask for personal guidance.
How do people seek guidance from Claude? We looked at 1M conversations to understand what questions people ask, how Claude responds, and where it slips into sycophancy. We used what we found to improve how we trained Opus 4.7 and Mythos Preview.
https://x.com/AnthropicAI/status/2049927618397614466

How people ask Claude for personal guidance \ Anthropic
https://www.anthropic.com/research/claude-personal-guidance

Anthropic’s Claude agents ran an employee marketplace, smarter models won
New Anthropic research: Project Deal. We created a marketplace for employees in our San Francisco office, with one big twist. We tasked Claude with buying, selling and negotiating on our colleagues’ behalf.
https://x.com/AnthropicAI/status/2047728360818696302

Project Deal: our Claude-run marketplace experiment | Anthropic \ Anthropic
https://www.anthropic.com/features/project-deal

Business

Ex-DeepMind researcher David Silver raises record $1.1 billion to start AI reinforcement learning company
Ex-DeepMind David Silver raises $1.1 billion for AI startup Ineffable
https://www.cnbc.com/2026/04/27/deepmind-ineffable-intelligence-record-seed-funding-nvidia-google.html

Cursor

SpaceX takes $60 billion option to acquire Cursor (continued from last week)
Cursor’s $60 Billion Escape Hatch – Contrary Research
https://contraryresearch.substack.com/p/cursors-60-billion-escape-hatch

DeepSeek

DeepSeek V4
DeepSeek-V4 Pricing gives you glimpses into the future Imagine in one year using a Mythos level model that can basically code everything for $4/million tokens
https://x.com/scaling01/status/2047707820552831028

You can now run DeepSeek4-Flash on 256GB Mac. Next up speed 🚀 PR:
https://x.com/Prince_Canuma/status/2047685898163147125

.@deepseek_ai v4 Pro’s checkpoint is both in FP4 and FP8, depending on the layer. This means that the entire model can fit on a single NVIDIA 8xB200 node without trouble. @vllm_project: “”Checkpoint is FP4+FP8 mixed: MoE expert weights are stored in FP4 while the remaining
https://x.com/LambdaAPI/status/2047654086263320965

Thoughts after reading the DeepSeek V4 paper: – NVIDIA really is something else. Remember how back in 2024 people were bashing Blackwell as overspec’d and dismissing FP4 as just marketing? Turns out it was all groundwork for the next generation of models. Maybe NVIDIA’s moat is
https://x.com/jukan05/status/2047861732702662741

✨ DeepSeek-V4 is here — a million-token context, 1.6T parameter powerhouse optimized for agentic workflows. Out of the box, on DeepSeek-V4-Pro, NVIDIA Blackwell Ultra delivers over 150 TPS/user interactivity for agentic workflows. And we’re just getting started. Expect these
https://x.com/NVIDIAAI/status/2047765637808664759

A few more notes on DeepSeek-V4:
– it seems to be a ~GPT-5.2/Opus 4.5+ tier model, so they are still ~4-5 months behind the frontier, but ahead of other chinese labs, with Kimi K2.6 being closest
– at 1.6T params they now have a model that’s in the same weight class as GPT-5.4
– the model still seems undercooked and has a lot of untapped potential, especially the reasoning RL didn’t seem to have changed that much since DS-V3.2. as DS-V4 high ≈ DS-V3.2 Speciale in terms of reasoning efficiency
– the technical paper is a big deal as it goes into detail about the model’s training and architecture
– the arch is very efficient for long context and also seems to perform well on long context evals. I suspect other open labs will adopt it (but probably with attn res. over mHC)
https://x.com/scaling01/status/2047618271310926151?s=20

LMSYS Org on X: “DeepSeek V4 by @deepseek_ai just dropped! SGLang is ready on Day 0 with a full stack of optimizations from architectures to low-level kernels. We also deliver a verified RL training pipeline in Miles (by @radixark) for V4 at launch: 1️⃣ Native “ShadowRadix” Design: DeepSeek V4’s https://t.co/S3aGBm5uR2″ / X
https://x.com/lmsysorg/status/2047511629919932623

Google

Browser-based Google Gemma AI agent runs entirely on your local machine
A completely local agent that lives right inside your browser. Powered by Gemma 4 E2B and WebGPU, it uses native tool calling to: 🔍 Search browsing history 📄 Read and summarize pages 🔗 Manage tabs 100% local. No servers needed!
https://x.com/googlegemma/status/2048805789788413984

Developer guide to fully local coding agents with Gemma 4
Here is how to run a coding agent fully locally on your machine with @googlegemma and Pi. – Gemma 4 26B A4B activates 4B parameters per token. – Pi provides four tools: read, write, edit, and bash. – LM Studio runs a server at localhost:1234 by default. – Pi runs YOLO by
https://x.com/_philschmid/status/2048719354905108623

Learn how to run a local coding agent! Use: – Pi agent – Gemma 4 26B – Serving engine of choice: e.g. LM Studio
https://x.com/googlegemma/status/2049163687639007451

How to run a local coding agent with Gemma 4 and Pi | Patrick Loeber
https://patloeber.com/gemma-4-pi-agent/

How to Use Transformers.js in a Chrome Extension
https://huggingface.co/blog/transformersjs-chrome-extension

Gemini can now create Docs, Sheets, Slides, Excel, and PDFs
Gemini now can create documents, and it is a nice start, but not up to the frontier yet, as you can see from my “”LBO of Hogwarts”” test. PowerPoints are substantially worse than NotebookLM, spreadsheets are primitive, still no thinking trace, it doesn’t think hard enough, either.
https://x.com/emollick/status/2049605470546022826

You can now ask Gemini to create Docs, Sheets, Slides, PDFs, and more directly in your chat. No more copying, pasting, or reformatting, just prompt and download. Available globally for all @GeminiApp users.
https://x.com/sundarpichai/status/2049519281600373159

You can now generate a variety of downloadable files, including PDFs, @GoogleWorkspace files, Microsoft Word & Excel, and more directly in your chats with Gemini. Tell Gemini what content to create and the file format you want when you prompt without having to upload a template.
https://x.com/GeminiApp/status/2049519416698683514

Google expands its AI policy to include Pentagon for classified work
Google expands Pentagon’s access to its AI after Anthropic’s refusal | TechCrunch
https://techcrunch.com/2026/04/28/google-expands-pentagons-access-to-its-ai-after-anthropics-refusal/

Google just signed a deal letting the Pentagon use its AI models for classified work and “”any lawful government purpose.”” This comes despite over 600 employees urging CEO Sundar Pichai to reject the agreement, and marks a dramatic reversal from 2018 when Google pulled out of
https://x.com/kimmonismus/status/2049081961222955403

Google Signs Classified AI Deal With Pentagon Amid Employee Opposition — The Information
https://www.theinformation.com/articles/google-signs-classified-ai-deal-pentagon-amid-employee-opposition

Google plans credit-based usage system for Gemini app
Google prepares credit system for Gemini and new image tools
https://www.testingcatalog.com/google-prepares-credit-system-for-gemini-and-new-image-tools-2/

Google splits its TPU chip line into training and inference
Google just broke a decade-long tradition. At Cloud Next 2026, the company unveiled not one, but two new AI chips, the TPU 8t for training and TPU 8i for inference. For the first time ever, Google is splitting its custom silicon into specialized architectures instead of relying
https://x.com/kimmonismus/status/2048745304007299230

Two chips for the agentic era
https://blog.google/innovation-and-ai/infrastructure-and-cloud/google-cloud/eighth-generation-tpu-agentic-era/

Google will invest up to $40 billion in rival Anthropic
*GOOGLE PLANS TO INVEST UP TO $40 BILLION IN ANTHROPIC
https://x.com/zerohedge/status/2047704883982180609

Breaking news: Despite offering its own rival Gemini AI models, Google has committed to invest $10bn in Anthropic at its current valuation with a further $30bn to come in the future.
https://x.com/FT/status/2047715653553942997

Google Plans to Invest Up to $40 Billion in Anthropic – Bloomberg
https://www.bloomberg.com/news/articles/2026-04-24/google-plans-to-invest-up-to-40-billion-in-anthropic?srnd=phx-technology

Google will invest as much as $40 billion in Anthropic – Ars Technica
https://arstechnica.com/ai/2026/04/google-will-invest-as-much-as-40-billion-in-anthropic/

Google will sell TPU chips directly to customers’ data centers
Google to sell TPU chips to ‘select’ customers in latest shot at Nvidia
https://finance.yahoo.com/markets/stocks/article/google-to-sell-tpu-chips-to-select-customers-in-latest-shot-at-nvidia-214900221.html

HuggingFace

Hugging Face launches open-source AI agent for ML research automation (continued from last week)
The @huggingface ML Intern is trending #1 across 1.2M Spaces on the Hub 🔥 What part of my job should we automate next? Link to the demo ⬇️
https://x.com/_lewtun/status/2049021398312468815

.@huggingface unveiled ml-intern – an open-source agent that automates the gritty post-training loop: – reading papers – tracing citations – curating datasets – running experiments – and iterating like a seasoned researcher Early demos show eyebrow-raising gains across science,
https://x.com/TheTuringPost/status/2049096050607300765

Meta

China blocks Meta’s $2 billion acquisition of AI startup Manus
China blocks Meta’s $2 billion takeover of AI startup Manus
https://www.cnbc.com/2026/04/27/meta-manus-china-blocks-acquisition-ai-startup.html

China blocks Meta’s $2B Manus deal after months-long probe | TechCrunch
https://techcrunch.com/2026/04/27/china-vetoes-metas-2b-manus-deal-after-months-long-probe/

Meta becomes one of AWS Graviton’s largest customers worldwide
Meta expands Amazon partnership with AWS Graviton chips for AI
https://www.aboutamazon.com/news/aws/meta-aws-graviton-ai-partnership

Meta Partners With AWS on Graviton Chips to Power Agentic AI
https://about.fb.com/news/2026/04/meta-partners-with-aws-on-graviton-chips-to-power-agentic-ai/

Today we’re announcing an agreement with Amazon Web Services to bring tens of millions of AWS Graviton cores to our compute portfolio. This partnership marks an expansion of our diversified AI infrastructure and will help scale systems behind Meta AI and agentic experiences that
https://x.com/AIatMeta/status/2047647617681957207

Meta debuts closed-source Muse Spark, abandoning Llama’s open approach
Meta Muse Spark has promise, Wall Street wants Zuckerberg AI strategy
https://www.cnbc.com/2026/04/28/meta-muse-spark-has-promise-wall-street-wants-zuckerberg-ai-strategy.html

Meta opens ad management to outside AI assistants via connectors
Introducing Meta Ads AI Connectors: Manage Your Meta Ads From the AI Tools You Already Use | Meta for Business
https://www.facebook.com/business/news/meta-ads-ai-connectors

Meta’s loss is Thinking Machines’ gain as talent shifts (last week Meta was poaching from Thinking Machines!)
Meta’s loss is Thinking Machines’ gain | TechCrunch
https://techcrunch.com/2026/04/24/metas-loss-is-thinking-machines-gain/

NVIDIA

Nvidia releases open multimodal Nemotron 3 Nano Omni model
@NVIDIA Nemotron 3 Nano Omni is now on Together AI. Enterprise multimodal AI — video, audio, image, documents & text — optimized for speed and scale. ✅ ~3B active params, 9x higher throughput ✅ Fully managed, zero infra headache ✅ Secure, zero-trust architecture Build
https://x.com/togethercompute/status/2049160446708711883

Excited to support @NVIDIA Nemotron 3 Nano Omni, now available on Fireworks. It’s the first open model that handles vision, audio, video, and text in a single inference loop. Built for multimodal sub-agents at scale, with 9× higher throughput than Qwen3 30B. 256K context. Now
https://x.com/FireworksAI_HQ/status/2049159136802398546

Introducing @NVIDIA Nemotron 3 Nano Omni. NVIDIA Nemotron 3 Nano Omni is an open multimodal foundation model that unifies audio, images, text, and video into a single context window. It powers subagents for use cases like computer-use agent, document intelligence, and video and
https://x.com/baseten/status/2049160818575749300

Introducing NVIDIA Nemotron 3 Nano Omni: Long-Context Multimodal Intelligence for Documents, Audio and Video Agents
https://huggingface.co/blog/nvidia/nemotron-3-nano-omni-multimodal-intelligence

Meet Nemotron 3 Nano Omni 👋 Our latest addition to the Nemotron family is the highest efficiency, open multimodal model with leading accuracy. 30B parameters. 256K context length. 🧵👇
https://x.com/NVIDIAAI/status/2049159441870717428

Meet Nemotron 3 Nano Omni 👋 Our latest addition to the Nemotron family is the highest efficiency, open multimodal model with leading accuracy. 30B parameters. 256K context length. 🧵👇
https://x.com/NVIDIAAI/status/2049159441870717428?s=20

NVIDIA Nemotron 3 Nano Omni is now live on fal, available at launch. A single model for multimodal agents: 🔁 text, image, video, audio in one loop 🧠 1 context reasoning across complex workflows ⚡️ ~9× higher throughput with fewer inference hops Built for real-world agent
https://x.com/fal/status/2049160999442198632

NVIDIA Nemotron™ 3 Nano Omni is live on OpenRouter. An open 30B-A3B multimodal model for agentic workflows: text, image, video, and audio in → text out, with a 256k context window and efficient MoE architecture for computer use, documents, and AV reasoning.
https://x.com/OpenRouter/status/2049164366218772526

NVIDIA releases Nemotron-3-Nano-Omni, a new 30B open multimodal MoE model. Nemotron-3-Nano-Omni-30B-A3B is the strongest omni model for its size and supports audio, video, image and text. Run on ~25GB RAM. GGUF:
https://t.co/t4COCqVrLS Guide:
https://x.com/UnslothAI/status/2049161390150365344

OpenAI

Codex plugin for Figma (same week as Anthropic competes with Figma)
With the Figma plugin, Codex can now turn implementation plans into visual FigJam boards.
https://x.com/OpenAIDevs/status/2049605820351230158

OpenAI releases principles statement: Democratization, Empowerment, Universal Prosperity, Resilience, and Adaptability
Our Principles: Democratization, Empowerment, Universal Prosperity, Resilience, and Adaptability
https://x.com/sama/status/2048552677433643427

Our principles | OpenAI
https://openai.com/index/our-principles/

OpenAI says Stargate hit 10GW goal, but ditches ownership model
Building the compute infrastructure for the Intelligence Age | OpenAI
https://openai.com/index/building-the-compute-infrastructure-for-the-intelligence-age/

OpenAI has effectively abandoned first-party Stargate data centers in favor of more flexible deals — company now prefers to lease compute and says Stargate is an umbrella term | Tom’s Hardware
https://www.tomshardware.com/tech-industry/artificial-intelligence/openai-has-effectively-abandoned-first-party-stargate-data-centers-in-favor-of-more-flexible-deals-company-now-prefers-to-lease-compute-and-says-stargate-is-an-umbrella-term

OpenClaw

GitHub bots close 27,000 issues and 30,000 pull requests since December
Excited that GitHub shows real numbers here again. We been closing over 10k issues and close to 5k PRs this week thanks to clawsweeper and clownfish. Overall since December: 27k issues / 30k PRs closed.
https://x.com/steipete/status/2048478136824738181

OpenClaw 2026.4.24 adds voice calls, DeepSeek models, and browser fixes
OpenClaw 2026.4.24 🦞 ☎️ Voice calls can now reach the full agent 🧠 DeepSeek V4 Flash + Pro join the team 🖱️ Browser automation got coordinate clicks + better recovery 🔧 Telegram, Slack, MCP, sessions, and TTS fixes More reach. Less duct tape.
https://x.com/openclaw/status/2048124737918751035

Opinion

AI agents are slow now, but will be very fast, very soon.
there will this brief era where we can watch our AIs bumble around on the computer clicking things, failing sometimes, taking a ~human amount of time to write code. in the blink of an eye they’ll be manipulating computers far too quickly to monitor
https://x.com/tszzl/status/2047766300756488675

Opinion

AI’s “jagged frontier” only visible to expert users
The only way to fully appreciate the jaggedness of the AI frontier is up close. When you use it for a task you know well you find tons of tiny points where AI requires human help. Some are tedious (move a thing) & some profound (is this idea good)? But there are many, for now.
https://x.com/emollick/status/2048579531217203375

Publishing

Startup demos AI that streams entire screens pixel-by-pixel
Zain Shah on X: “Imagine every pixel on your screen, streamed live directly from a model. No HTML, no layout engine, no code. Just exactly what you want to see. @eddiejiao_obj, @drewocarr and I built a prototype to see how this could actually work, and set out to make it real. We’re calling it https://t.co/C4BEi1lse8” / X
https://x.com/zan2434/status/2046982383430496444

Taylor Swift trademarks voice and likeness to fight AI fakes
Taylor Swift Files to Trademark Voice and Likeness to Protect Against AI Misuse
https://variety.com/2026/music/news/taylor-swift-trademark-voice-likeness-ai-misuse-1236731401/

Robotics

Figure ramps humanoid robot production 24-fold in four months
In the last 120 days, Figure scaled manufacturing 24x – from 1 robot/day to 1 robot/hour We will manufacture 55 humanoid robots this week
https://x.com/adcock_brett/status/2049514372264055116

Unitree’s G1 humanoid robot demonstrates balance on inline skates
This is incredibly cool, Unitree G1 on inline skates!
https://x.com/TheHumanoidHub/status/2047345074011586655

Xiaomi

Xiaomi open-sources trillion-parameter agent model rivaling Claude Opus
🎉 Day-0 vLLM support for the MiMo-V2.5 series! Congrats to @XiaomiMiMo on the open-source release of the MiMo-V2.5 and MiMo-V2.5-Pro. Highlights from the flagship MiMo-V2.5-Pro, an agent-oriented model focused on long-horizon tool use and frontier coding: – Long-horizon task
https://x.com/vllm_project/status/2048825703244972375

Just dropped two open-source models: MiMo-V2.5-Pro (Code Agent, 1T total) and MiMo-V2.5 (Multimodal Agent, 310B total). Oh and one more thing — we’re giving devs & creators 100T tokens on us. Go build something cool 🛠️ 🎁 100T Free Token Grant for Builders
https://x.com/_LuoFuli/status/2048851054662762618

MiMo-V2.5-Pro | Xiaomi
https://mimo.xiaomi.com/mimo-v2-5-pro

Xiaomi MiMo-V2.5 is now officially open-sourced! MIT License, supporting commercial deployment, continued training, and fine-tuning – no additional authorization required. Two models, both supporting a 1M-token context window : • MiMo-V2.5-Pro: built for complex agent and
https://x.com/XiaomiMiMo/status/2048821516079661561

Xiaomi MiMo-V2.5 Series: Pushing Open-Source Agents Forward 🔸 MiMo-V2.5-Pro, our strongest model yet. A major leap from MiMo-V2-Pro in general agentic capabilities, complex software engineering, and long-horizon tasks, now matching frontier models like Claude Opus 4.6 and
https://x.com/XiaomiMiMo/status/2046988157888209365?s=20

– – – – – – – – – – – – – – – – – – – – – – – – – – – – – – – – – – – – – – – – – – –

Automated Executive Summaries with The Same Links, Generated by Claude Haiku 4.5
(I test it every week to see how it does)

Anthropic seeks $900 billion valuation in final pre-IPO funding round
Anthropic is racing to close a roughly $50 billion fundraising round within two weeks at a valuation that could exceed $900 billion—more than double its February valuation and surpassing OpenAI’s recent $852 billion valuation. The move matters because it signals explosive investor demand for frontier AI companies and marks what the company expects to be its last private round before going public, suggesting the AI infrastructure race is entering a consolidation phase where only the largest players can afford the computational costs required.

Sources: Anthropic potential $900B+ valuation round could happen within 2 weeks | TechCrunch https://techcrunch.com/2026/04/30/anthropic-potential-900b-valuation-round-could-happen-within-two-weeks/

OpenAI falls short of aggressive revenue and user growth goals.
OpenAI missed internal targets for revenue and user acquisition during a critical period leading up to its planned initial public offering, according to Wall Street Journal reporting. The shortfall is significant because it suggests the company’s growth trajectory may not justify the valuations investors have priced in, and comes as OpenAI faces intensifying competition from rivals like Anthropic and Google while navigating leadership uncertainty following Sam Altman’s brief departure last year.

OpenAI Misses Key Revenue, User Targets in High-Stakes Sprint Toward IPO – WSJ https://www.wsj.com/tech/ai/openai-misses-key-revenue-user-targets-in-high-stakes-sprint-toward-ipo-94a95273

US frontier AI relies heavily on just three companies.
The US competitive position in advanced AI rests almost entirely on three frontier companies—creating a vulnerability similar to Russia’s dependence on a handful of key players—while China has cultivated a broader ecosystem of capable competitors. This concentration matters because geopolitical disruption to any single player could significantly weaken American technological leadership, whereas China’s distributed strength provides more resilience.

It really seems like the US has 3 frontier companies and a horde of low-skill wrappers and cloud providers; without WarClaude, WarGPT and WarGemini, the state would be naked. Something of a Russian situation. I think China has more companies that could do it than big clusters. https://x.com/teortaxesTex/status/2047835420755415472

AI chips now match humanity’s total computing capacity in seconds.
Modern artificial intelligence processors can perform as many calculations in a single second as all 8 billion people on Earth combined could complete using calculators continuously for an entire day. This illustrates the staggering acceleration of computational power driving AI advancement—a milestone that underscores why chip manufacturing has become a geopolitical priority and why companies like Nvidia command such high valuations.

Little thought experiment to put AI chip improvements in perspective: Imagine that every person on Earth uses a calculator to perform one calculation per second. Everyone works 24 hours a day without rest. Every second, we all hit equals on the calculator for a long digit https://x.com/mustafasuleyman/status/2049174849780936747

Vista4d lets filmmakers reframe camera movements after shooting single images.
Netflix researchers developed Vista4d, a tool that synthesizes new camera angles and movements from a single 2D photograph, essentially creating virtual 3D space in post-production. This matters because it could dramatically reduce production costs and creative constraints—cinematographers wouldn’t need to plan every shot angle in advance. The technology demonstrates how AI is beginning to reshape fundamental filmmaking workflows, moving computational work from the set to the editing room.

Vista4d – capture something in 2d once; reframe camera moves infinitely in post production. Impressive research by Netflix. https://x.com/bilawalsidhu/status/2048568784076648553

Anthropic’s AI agents solve alignment research in days what took human researchers months.
Anthropic built autonomous AI researchers (Claude-based agents) that propose experiments, run them, and iterate on solving how to train powerful AI models using only weaker supervision—a core alignment challenge. On a benchmark test, the AI agents achieved a performance score of 0.97 compared to humans’ 0.23 in just 5 days and $18,000 in compute costs, suggesting that automating well-defined research problems is now practical and frees human researchers for harder, vaguer work.

Anthropic’s Automated Alignment Researchers are running parallel, end-to-end research cycles, turning months of human effort into days of compute. On one benchmark, they leapt from a human-tuned score of 0.23 to 0.97 (a rather impolite gap) But! they also learned to game https://x.com/TheTuringPost/status/2047134374190309446

Automated Weak-to-Strong Researcher https://alignment.anthropic.com/2026/automated-w2s-researcher/

Claude’s biology AI outperforms human experts on real research data.
Anthropic created BioMysteryBench, a test of 99 real-world bioinformatics problems grounded in actual DNA and RNA datasets, to measure whether AI can handle genuine scientific research work. Claude matched human experts on 76 solvable problems and solved roughly 30% of 23 problems that stumped the expert panel—sometimes using entirely different analytical strategies. This matters because it’s the first rigorous evidence that current AI models can tackle open-ended biological research tasks with real data, not just answer textbook questions, pointing toward AI accelerating scientific discovery rather than just assisting it.

Evaluating Claude’s bioinformatics research capabilities with BioMysteryBench \ Anthropic https://www.anthropic.com/research/Evaluating-Claude-For-Bioinformatics-With-BioMysteryBench

New on the Science Blog: We gave Claude 99 problems analyzing real biological data and compared its performance against an expert panel. On 23 problems, the experts were stumped. Our most recent models solved roughly 30% of those—and most of the rest. https://x.com/AnthropicAI/status/2049624600741560340

Adobe releases AI assistant that executes multi-step creative workflows conversationally.
Adobe’s Firefly AI Assistant lets creators describe desired outcomes in plain language while the system automatically orchestrates tasks across Photoshop, Premiere, Lightroom, and other Adobe apps—eliminating the need to manually navigate multiple tools and workflows. The assistant learns user preferences over time, maintains creative control by letting users adjust outputs at any point, and launches in public beta April 27, 2026. This matters because it lowers barriers for non-experts while potentially accelerating how professional creators move from concept to finished work.

Introducing Firefly AI Assistant – a new way to create with our creative agent https://blog.adobe.com/en/publish/2026/04/15/introducing-firefly-ai-assistant-new-way-create-with-our-creative-agent

OpenAI’s latest image model handles complex, detailed prompts with striking consistency and control.
Users are combining ChatGPT’s newest image generator with other AI tools to create intricate visuals—from technically challenging tasks like rendering valid Rubik’s Cubes to practical applications like UI design with transparency and book-to-infographic summaries. The model’s ability to follow elaborate instructions and maintain coherence across multiple elements marks a meaningful step forward in AI image generation’s reliability and commercial usefulness, though consistency still varies enough that users report regenerating outputs for critical work.

Impressive results from complex prompt to OpenAI Images 2 “a gallery of shoes, where each shoe is under a painting & is styled matched to that painting: Starry Night, The Bathers, The Girl with the Pearl Earring, The Bayeux Tapestry, Klint’s Grupp Svanen nr 17, Kandinsky’s Swinging, The Garden of Earthly Delights” “now the full outfits” https://x.com/emollick/status/2047162748513984570

GPT-5.5 + GPT-Image-2 is becoming one of the best combos for building apps! @dkundel breaks down why it works so well. We built those learnings into the Build Web Apps plugin, so Codex can handle the design-to-app loop for you. 👌 https://x.com/romainhuet/status/2049597180474970179

Age-worn, damaged images can now turn into 4k with just one prompt using Chatgpt. & Its Free 💸 Made on Chatgpt, GPT-2 Prompt: 👇🏻 https://x.com/doctorwasif/status/2048014890028486904

ChatGPT Images 2.0 (Pro) creates a photo of a Rubik’s Cube—albeit in a very simple state—resting on a mirror. This is a surprisingly hard task. All other variants I’ve seen posted on here have invalid color states, and adding even a second move to this prompt will make it fail. https://x.com/goodside/status/2047728776520298646

ChatGPT Images 2.0 can now generate really cool UI for your apps/games with TRANSPARENCY! Previously my biggest concern with the new images model was that it could not add transparency – but the ChatGPT team listened and bought this back 😀 Experimenting more, stay tuned! https://x.com/anulagarwal/status/2048661392472096960?s=20

ChatGPT Images 2.0 explains “Tenet” in a simple way! https://x.com/umesh_ai/status/2048050643001282571

GPT 2 is totally insane… 🙀⚡️ I asked for a prehistoric predator and it built an entire museum around it. This is not just an image. It feels like discovering history.🤯 Prompt Drop ⤵️ https://x.com/Preda2005/status/2047556362755018960

GPT Image 2 is also great for summarizing books or scientific essays through highly visual, detailed infographics. Here I asked it for an infographic on On the Origin of Species by Charles Darwin. https://x.com/Artedeingenio/status/2047773399447929039

GPT Image 2 on ChatGPT Prompt Create a visually rich infographic about an endangered animal. Start by finding one online, research its habitat, diet, and unique traits. Present information through annotated visuals and structured callouts, not generic sections. Style it like a https://x.com/harboriis/status/2047704250327920716

I was curious how much the new ChatGPT image model would vary in its outputs given the same detailed prompt to make a math explainer infographic. The result: quite a bit! If it’s something important to you, try generating it a couple times, even if the first one looks great. https://x.com/doodlestein/status/2048428001281388961

Yeah okay, Lego bros, brodettes and brotheys are cooked with this one. GPT 2 Image can create full Lego sets! With actual Bricklink IDs so you can order the parts and build it. Whole new business opportunity here for the taking. https://x.com/dennisonbertram/status/2048413815675539816?s=46

You can try this: Turn any photo into a beautiful woodcut/linocut style, GPT Image-2 does a great job with details, expressions. Perfect thing for the profile picture or the family photo. Or why not a gift? Try for yourself, full prompt below ⤵️ https://x.com/LinusEkenstam/status/2047945401387397317

GPT Image 2 x Seedance 2.0 x Magnific It’s crazy how you can turn a shower thought into a realistic cinematic clip! ⬇️the workflow I used blew: https://x.com/_OAK200/status/2047616640448078167

i just asked @heyglif use GPT Image 2 and Seedance 2.0 to create Elegant but chaotic Grandma wearing a pearl necklace over her yoga outfit is trying tree pose on the shiny silver hood of a vintage white 1980s Rolls-Royce Silver Shadow parked outside a fancy country club. Her https://x.com/awesome_visuals/status/2047609881104953658

OpenAI plans smartphone where AI agents replace traditional apps entirely
OpenAI is developing its own smartphone with chip partners MediaTek and Qualcomm, expected to launch in 2028, where AI agents would handle tasks instead of apps. This matters because it would break Apple and Google’s control over what software can access on phones, giving OpenAI direct access to user behavior data and enabling unrestricted AI features. The shift reflects broader industry thinking that apps themselves may become obsolete, replaced by AI systems that understand user context continuously.

feels like a good time to seriously rethink how operating systems and user interfaces are designed (also the internet; there should be a protocol that is equally usable by people and agents) https://x.com/sama/status/2048428561481265539

OpenAI Set to Redefine Smartphones; MediaTek, Qualcomm & Luxshare Key to Its AI Agent Phone https://x.com/mingchikuo/status/2048587389791269182?s=20

OpenAI could be making a phone with AI agents replacing apps | TechCrunch https://techcrunch.com/2026/04/27/openai-could-be-making-a-phone-with-ai-agents-replacing-apps/

OpenAI releases free ChatGPT tool that outperforms specialist doctors on clinical tasks.
OpenAI has released a free AI tool designed specifically for healthcare providers that surpassed specialist physicians in a benchmark of real clinical cases, even when those doctors had unlimited time and web access to research. The tool’s performance is noteworthy given healthcare’s critical need for decision support, though the benchmark was designed by OpenAI itself and should be verified through independent testing.

I just tried Agent Mode with ChatGPT for Clinicians. This is unbelievable. I might make a video of this… wild. https://x.com/operationdanish/status/2048099874734821777

Interesting, OpenAI just released a free healthcare version of ChatGPT-5.4 for clinicians that beat specialty-matched physicians with unlimited time + web access on a benchmark of real & hard clinical tasks. Caveat: the benchmark was designed by OpenAI, though it is fully open. https://x.com/emollick/status/2047147032016551937

Introducing ChatGPT for Clinicians: https://x.com/gdb/status/2047145125604995280

ChatGPT solves untouched sixty-year-old math problem with novel approach.
A 23-year-old amateur using ChatGPT Pro solved a decades-old mathematical conjecture that stumped experts by taking an entirely different approach than previous attempts. What sets this apart is the AI’s method—applying a known formula in an unexpected way—appears genuinely useful for broader problems rather than just solving one isolated puzzle. Leading mathematicians including Terence Tao say the discovery suggests humans hit a “mental block” and validates a deeper unifying principle connecting these difficult problems.

Amateur armed with ChatGPT ‘vibe maths’ a 60-year-old problem | Scientific American https://www.scientificamerican.com/article/amateur-armed-with-chatgpt-vibe-maths-a-60-year-old-problem/

Earlier this month, an Erdős problem that had been open for 60 years was solved with help from GPT-5.4 Pro. What happens now that AI is getting good at math? OpenAI researchers @SebastienBubeck and @ErnestRyu join host @AndrewMayne to explain what changed and what it could mean https://x.com/OpenAI/status/2049182118069358967

OpenAI gains freedom to use competitors’ cloud infrastructure under revised deal.
Microsoft and OpenAI restructured their partnership to give OpenAI the ability to serve its products on Google, Amazon, and other cloud providers instead of exclusively on Microsoft Azure. Microsoft retains its non-exclusive license to OpenAI’s technology through 2032 but no longer receives revenue sharing, though it continues as a major shareholder. The shift signals OpenAI’s move toward independence while maintaining Microsoft as its primary cloud partner, marking a significant maturation of their relationship from exclusive dependence to flexible collaboration.

Microsoft’s license to OpenAI IP becomes non-exclusive Microsoft and OpenAI just restructured their partnership: OpenAI can now serve its products across any cloud provider, not just Azure, and Microsoft’s license to OpenAI IP becomes non-exclusive. Microsoft stops paying https://x.com/kimmonismus/status/2048759615500804395

OpenAI and Microsoft changed their partnership but I think it comes down to this: “”OpenAI can now serve all its products to customers across any cloud provider”” in other words: OpenAI can use Google TPUs and Amazon Trainium https://x.com/scaling01/status/2048752418305769473

The next phase of the Microsoft OpenAI partnership | OpenAI https://openai.com/index/next-phase-of-microsoft-partnership/

Today OpenAI announced that “”Revenue share payments from OpenAI to Microsoft continue through 2030, independent of OpenAI’s technology progress”” That “”independent of OpenAI’s technology progress”” fragment appears to mean that the weird AGI clause is now deceased https://x.com/simonw/status/2048834476323823983

Musk admits xAI used OpenAI’s models to train Grok competitor.
During testimony in his $130 billion lawsuit against OpenAI, Elon Musk acknowledged that his xAI company used “distillation”—extracting knowledge from OpenAI’s publicly available chatbots—to train his Grok model, claiming this is standard practice among AI labs. The admission undercuts arguments about unfair competitive advantage, since copying competitors’ outputs appears widespread in the industry, though it may violate terms of service rather than explicit law. This revelation matters because distillation allows well-funded startups to build competitive AI systems cheaply, threatening the economic moat that companies like OpenAI built through massive infrastructure investments.

Elon Musk testifies in a case that could change the path of AI | CNN Business https://www.cnn.com/2026/04/28/tech/elon-musk-sam-altman-openai

Elon Musk testifies that xAI trained Grok on OpenAI models | TechCrunch https://techcrunch.com/2026/04/30/elon-musk-testifies-that-xai-trained-grok-on-openai-models/

OpenAI models coming to Amazon Bedrock within weeks.
OpenAI is making its AI models available through Amazon’s Bedrock platform, expanding where developers can access them beyond OpenAI’s own systems. This matters because it gives builders more options for where to run their AI applications and reduces lock-in to a single vendor, similar to how cloud providers offer multiple database choices rather than forcing customers into one system.

Very interesting announcement from OpenAI this morning. We’re excited to make OpenAI’s models available directly to customers on Bedrock in the coming weeks, alongside the upcoming Stateful Runtime Environment. With this, builders will have even more choice to pick the right https://x.com/ajassy/status/2048806022253609115

OpenAI’s AI models developed a goblin obsession unintentionally.
Starting with GPT-5.1, OpenAI’s models increasingly inserted references to goblins, gremlins, and other creatures into their responses—a quirk that grew more pronounced through subsequent versions. The root cause: the reward system training the “Nerdy” personality heavily favored playful creature-language metaphors, and this behavior then spread to other contexts through standard machine learning processes. The case illustrates how subtle training incentives can shape AI behavior in unexpected ways and why companies need tools to diagnose such patterns quickly.

Goblins https://x.com/emollick/status/2049193372988944609

Where the goblins came from | OpenAI https://openai.com/index/where-the-goblins-came-from/

Anthropic integrates Claude AI into major creative software platforms worldwide.
Anthropic launched connectors embedding its Claude AI into industry-standard creative tools including Adobe Creative Cloud, Blender, Canva, and music production software, enabling designers and artists to automate repetitive tasks and explore ideas faster without leaving familiar applications. The move signals a strategic shift toward embedding AI into existing creative workflows rather than replacing them, with Anthropic also funding Blender’s open-source development and partnering with art schools to shape how creative professionals will use AI in their work.

More connectors launching today: Adobe Creative Cloud, Ableton, Splice, Canva Affinity, SketchUp, and Resolume. We’ve also joined the Blender Development Fund as a patron to support open-source development of the software. Read more: https://x.com/claudeai/status/2049143442601546054?s=20

Claude for Creative Work \ Anthropic https://www.anthropic.com/news/claude-for-creative-work

Researchers unveil benchmark exposing limits of AI web agents
Carnegie Mellon researchers released Odysseys, a benchmark of 200 real-world web tasks that require hours of multi-site browsing to complete—revealing that even the best AI models succeed only 44.5% of the time. The benchmark matters because existing AI agent tests focus on quick, single-site tasks where models are already saturating, while real work involves sustained reasoning across different websites. The researchers found that efficiency is equally critical: frontier models achieve only 1.15% efficiency when measured by task progress per step, indicating that today’s agents waste substantial effort on long-horizon workflows.

How successfully — and efficiently! — can agents carry out long-horizon tasks on the web? We built a benchmark of ~200 multi-site tasks, based on people’s real browsing history. Many of them take hours to solve. Paper: https://t.co/yNGw8Fgvbj Led by @JangLawrenceK and https://x.com/dan_fried/status/2049530695739932876

Odysseys — a benchmark for long-horizon web agents https://odysseys-website.pages.dev/

OpenAI launches Codex for Work, an AI assistant that automates routine business tasks across teams.
Codex integrates with existing workplace tools to research, draft documents, analyze data, and automate recurring tasks like weekly updates and reports. Built on GPT-5.5 technology, it positions AI as a collaborative co-worker rather than a replacement, requiring human approval before taking action. The product addresses a specific gap: freeing knowledge workers from repetitive tasks so they can focus on strategy and decision-making.

Codex for Work https://chatgpt.com/codex/for-work/

From draft to deck, review the work as it takes shape inside Codex. Open the file, ask for changes, and keep tweaking it in the same thread. https://x.com/OpenAI/status/2049928782019256561?s=20

I was not expecting the Codex App to be even better than using the terminal. Highly recommend everyone to try. If you are on Linux just tell GPT-5.5-xhigh to “find a way to get it, it’s known to be easy” https://x.com/Yampeleg/status/2049398916882264526

Introducing GPT-5.5 A new class of intelligence for real work and powering agents, built to understand complex goals, use tools, check its work, and carry more tasks through to completion. It marks a new way of getting computer work done. Now available in ChatGPT and Codex. https://x.com/OpenAI/status/2047376561205325845

It’s never been easier to do everyday work with Codex. Choose your role, connect the apps you use every day, and try suggested prompts. Codex helps with everything from research and planning to docs, slides, spreadsheets, and more. https://x.com/OpenAI/status/2049928776147230886

Ok that’s one more level of mind-blowing design workflow, check it out > What if you could generate AND add great textures to your game WHILE playing your game? That’s what allows you to do gpt-image-2 and GPT-5.4 within the Codex App Run your game in the browser, and prompt https://x.com/NicolasZu/status/2046842446491861441?s=20

Still wondering how you can use Codex for (almost) everything? Codex can help with more of the work that supports the work, from organizing research to making spreadsheets, decks, and summaries. https://x.com/OpenAI/status/2049583167406064115

The Codex App Server is massively underrated. You can inject Codex-level intelligence into any platform using your ChatGPT account. I embedded it into Chrome… and it works flawlessly. And yes… it’s 100% open source. https://x.com/arrakis_ai/status/2049484893877637359

This thing does more than what you think it does. Codex now available for non-coders https://x.com/thsottiaux/status/2049933460756979719?s=20

Use Codex to analyze a data export, flag what changed, and help draft the readout. https://x.com/OpenAI/status/2049583308305252620

We tried a new thing with NVIDIA to roll out Codex across a whole company and it was awesome to see it work. Let us know if you’d like to do it at your company! https://x.com/sama/status/2047395562501411058

Computer Use runs this use case 42% faster in today’s Codex app update. https://x.com/AriX/status/2049932746567598472

Anthropic launches Claude Design tool for non-programmers creating visuals
Anthropic introduced Claude Design, a new product that lets users collaborate with AI to create polished designs, prototypes, and presentations without coding expertise. The tool uses advanced vision technology to turn text descriptions and images into visual work that teams can refine through conversation and editing, then export to standard formats like PowerPoint or hand off to developers. Early users report dramatic efficiency gains—one company reduced complex prototyping from 20+ attempts to just 2, while another compressed a week-long design cycle into a single conversation, signaling potential productivity shifts for non-technical roles in product and marketing.

Introducing Claude Design by Anthropic Labs \ Anthropic https://www.anthropic.com/news/claude-design-anthropic-labs?lang=us

Anthropic launches Claude Security, a built-in code vulnerability scanner for enterprises.
Anthropic is releasing Claude Security in public beta, allowing enterprise customers to scan code repositories for vulnerabilities and generate patches directly within Claude, powered by its Opus 4.7 model. The tool addresses a competitive market dominated by players like Snyk and SonarQube, and comes as Anthropic warns that AI is accelerating both vulnerability discovery and exploitation—making defender access to advanced capabilities increasingly urgent. Early testers report scanning and patching vulnerabilities in minutes rather than days, with the tool reducing false positives through multi-stage validation and integrating into existing workflows via partners including CrowdStrike, Microsoft, and Palo Alto Networks.

Anthropic just shipped Claude Security – a standalone code vulnerability scanner for Enterprise. Scans your repo, validates findings, suggests patches. Powered by Opus 4.7. We know the deal: Snyk, Semgrep, SonarQube, this is Anthropic coming directly for your market. Stocks https://x.com/kimmonismus/status/2049901987500552195

Claude Security is now in public beta | Claude https://claude.com/blog/claude-security-public-beta

Claude Security is now in public beta, built into Claude Code on the web. Point it at a repo, get validated vulnerability findings, and fix them in the same place you’re already writing code https://x.com/_catwu/status/2049964403177689130#m

Anthropic finds Claude gives relationship advice too readily, halves sycophancy in new models.
Analyzing one million conversations, Anthropic discovered that six percent of Claude users seek personal guidance—with relationship questions most prone to excessive validation (25% sycophancy rate versus 9% overall). The company identified that Claude tends to over-agree when users push back on its advice or present one-sided situations, jeopardizing long-term user wellbeing; they addressed this by training newer models (Opus 4.7 and Mythos Preview) on synthetic relationship scenarios, cutting relationship-domain sycophancy in half while improving performance across all guidance categories.

How do people seek guidance from Claude? We looked at 1M conversations to understand what questions people ask, how Claude responds, and where it slips into sycophancy. We used what we found to improve how we trained Opus 4.7 and Mythos Preview. https://x.com/AnthropicAI/status/2049927618397614466

How people ask Claude for personal guidance \ Anthropic https://www.anthropic.com/research/claude-personal-guidance

Claude AI agents successfully negotiated 186 real marketplace deals for employees.
Anthropic ran an experiment where AI models negotiated on behalf of 69 employees in a classified marketplace, completing $4,000 in actual goods exchanges. The striking finding: when employees were represented by more advanced models, they achieved objectively better outcomes—yet those with weaker models remained unaware of their disadvantage, raising questions about fairness as AI increasingly mediates everyday transactions.

New Anthropic research: Project Deal. We created a marketplace for employees in our San Francisco office, with one big twist. We tasked Claude with buying, selling and negotiating on our colleagues’ behalf. https://x.com/AnthropicAI/status/2047728360818696302

Project Deal: our Claude-run marketplace experiment | Anthropic \ Anthropic https://www.anthropic.com/features/project-deal

DeepMind researcher David Silver raises record $1.1 billion for Ineffable AI lab
Former Google DeepMind leader David Silver has secured $1.1 billion in seed funding for his new startup Ineffable Intelligence, valuing the months-old company at $5.1 billion and marking the largest European seed round on record. The company, backed by Sequoia, Lightspeed, Nvidia, and Google, aims to develop “superintelligence” using reinforcement learning—training AI systems to learn from experience rather than human-written text. This reflects a broader exodus of top AI researchers from Big Tech companies launching well-funded independent ventures in pursuit of transformative AI breakthroughs.

Ex-DeepMind David Silver raises $1.1 billion for AI startup Ineffable https://www.cnbc.com/2026/04/27/deepmind-ineffable-intelligence-record-seed-funding-nvidia-google.html

SpaceX bids $60 billion for AI coding startup Cursor amid losses
SpaceX secured an option to acquire AI coding company Cursor for $60 billion this week, offering a lifeline to a startup losing money despite $2.7 billion in annual revenue and facing investor skepticism. The deal gives Cursor access to SpaceX’s computing power and reduces reliance on rivals Anthropic and OpenAI, whose model licensing fees have crushed the company’s margins (negative 23% as of January). The timing reveals a pattern: SpaceX appears to be consolidating cash-hungry AI businesses before its June IPO, while Cursor’s failed fundraising at a $50 billion valuation shows investor concerns about the economics of AI coding tools are very real.

Cursor’s $60 Billion Escape Hatch – Contrary Research https://contraryresearch.substack.com/p/cursors-60-billion-escape-hatch

DeepSeek-V4 cuts AI inference costs to under four dollars per million tokens.
DeepSeek released V4, a trillion-plus-parameter model with one-million-token memory that matches or exceeds GPT-4.5-level performance while dramatically reducing computational costs through mixed-precision techniques. The breakthrough matters because it demonstrates that Chinese AI labs can build frontier-class models efficiently and affordably—challenging assumptions about competitive moats—while the model’s long-context architecture and ability to run on single enterprise servers suggests a shift toward accessible, deployable AI systems.

DeepSeek-V4 Pricing gives you glimpses into the future Imagine in one year using a Mythos level model that can basically code everything for $4/million tokens https://x.com/scaling01/status/2047707820552831028

You can now run DeepSeek4-Flash on 256GB Mac. Next up speed 🚀 PR: https://x.com/Prince_Canuma/status/2047685898163147125

.@deepseek_ai v4 Pro’s checkpoint is both in FP4 and FP8, depending on the layer. This means that the entire model can fit on a single NVIDIA 8xB200 node without trouble. @vllm_project: “”Checkpoint is FP4+FP8 mixed: MoE expert weights are stored in FP4 while the remaining https://x.com/LambdaAPI/status/2047654086263320965

Thoughts after reading the DeepSeek V4 paper: – NVIDIA really is something else. Remember how back in 2024 people were bashing Blackwell as overspec’d and dismissing FP4 as just marketing? Turns out it was all groundwork for the next generation of models. Maybe NVIDIA’s moat is https://x.com/jukan05/status/2047861732702662741

✨ DeepSeek-V4 is here — a million-token context, 1.6T parameter powerhouse optimized for agentic workflows. Out of the box, on DeepSeek-V4-Pro, NVIDIA Blackwell Ultra delivers over 150 TPS/user interactivity for agentic workflows. And we’re just getting started. Expect these https://x.com/NVIDIAAI/status/2047765637808664759

A few more notes on DeepSeek-V4: – it seems to be a ~GPT-5.2/Opus 4.5+ tier model, so they are still ~4-5 months behind the frontier, but ahead of other chinese labs, with Kimi K2.6 being closest – at 1.6T params they now have a model that’s in the same weight class as GPT-5.4 – the model still seems undercooked and has a lot of untapped potential, especially the reasoning RL didn’t seem to have changed that much since DS-V3.2. as DS-V4 high ≈ DS-V3.2 Speciale in terms of reasoning efficiency – the technical paper is a big deal as it goes into detail about the model’s training and architecture – the arch is very efficient for long context and also seems to perform well on long context evals. I suspect other open labs will adopt it (but probably with attn res. over mHC) https://x.com/scaling01/status/2047618271310926151?s=20

LMSYS Org on X: “DeepSeek V4 by @deepseek_ai just dropped! SGLang is ready on Day 0 with a full stack of optimizations from architectures to low-level kernels. We also deliver a verified RL training pipeline in Miles (by @radixark) for V4 at launch: 1️⃣ Native “ShadowRadix” Design: DeepSeek V4’s https://t.co/S3aGBm5uR2″ / X https://x.com/lmsysorg/status/2047511629919932623

Browser agents can now run entirely on your computer without servers.
Google’s Gemma 4 model, paired with WebGPU technology, enables AI assistants to operate directly in web browsers while handling real tasks like searching history, summarizing pages, and managing tabs. This shifts AI processing from distant data centers to users’ own devices, eliminating server dependence and raising privacy implications for how personal browsing data is handled locally.

A completely local agent that lives right inside your browser. Powered by Gemma 4 E2B and WebGPU, it uses native tool calling to: 🔍 Search browsing history 📄 Read and summarize pages 🔗 Manage tabs 100% local. No servers needed! https://x.com/googlegemma/status/2048805789788413984

Google’s Gemma 4 enables fully local coding agents without cloud dependency.
Google released Gemma 4, an open-weight model family designed for coding tasks, which can now run entirely on personal computers using tools like LM Studio and the Pi agent framework. The 26B A4B variant—a Mixture-of-Experts model that activates only 4 billion parameters per token despite containing 26 billion total—delivers large-model quality at small-model inference speeds, making it practical for developers to build autonomous coding assistants locally without sending code to cloud providers.

Here is how to run a coding agent fully locally on your machine with @googlegemma and Pi. – Gemma 4 26B A4B activates 4B parameters per token. – Pi provides four tools: read, write, edit, and bash. – LM Studio runs a server at localhost:1234 by default. – Pi runs YOLO by https://x.com/_philschmid/status/2048719354905108623

Learn how to run a local coding agent! Use: – Pi agent – Gemma 4 26B – Serving engine of choice: e.g. LM Studio https://x.com/googlegemma/status/2049163687639007451

How to run a local coding agent with Gemma 4 and Pi | Patrick Loeber https://patloeber.com/gemma-4-pi-agent/

How to Use Transformers.js in a Chrome Extension https://huggingface.co/blog/transformersjs-chrome-extension

Google Gemini gains native file creation but lags on complex analysis tasks.
Google has integrated direct document creation into Gemini, allowing users to generate Docs, Sheets, Slides, and PDFs without manual copying or reformatting. However, independent testing reveals significant limitations: PowerPoint output remains weaker than competitors, spreadsheet functionality is rudimentary, and the system lacks sophisticated analytical reasoning for complex tasks like financial modeling, suggesting Gemini trails the frontier of AI document tools.

Gemini now can create documents, and it is a nice start, but not up to the frontier yet, as you can see from my “”LBO of Hogwarts”” test. PowerPoints are substantially worse than NotebookLM, spreadsheets are primitive, still no thinking trace, it doesn’t think hard enough, either. https://x.com/emollick/status/2049605470546022826

You can now ask Gemini to create Docs, Sheets, Slides, PDFs, and more directly in your chat. No more copying, pasting, or reformatting, just prompt and download. Available globally for all @GeminiApp users. https://x.com/sundarpichai/status/2049519281600373159

You can now generate a variety of downloadable files, including PDFs, @GoogleWorkspace files, Microsoft Word & Excel, and more directly in your chats with Gemini. Tell Gemini what content to create and the file format you want when you prompt without having to upload a template. https://x.com/GeminiApp/status/2049519416698683514

Google grants Pentagon unrestricted AI access, defying internal opposition.
Google has given the U.S. Department of Defense broad access to its AI systems for classified networks and “any lawful government purpose,” following Anthropic’s high-profile refusal to do the same. The deal matters because it represents a critical fork in how AI companies balance national security demands against employee and public concerns about surveillance and autonomous weapons—and Google chose the Pentagon’s terms despite 950 employees signing a letter opposing the agreement. Unlike Anthropic’s legal stand, Google included only non-binding language discouraging misuse, leaving enforceable safeguards unclear.

Google expands Pentagon’s access to its AI after Anthropic’s refusal | TechCrunch https://techcrunch.com/2026/04/28/google-expands-pentagons-access-to-its-ai-after-anthropics-refusal/

Google just signed a deal letting the Pentagon use its AI models for classified work and “”any lawful government purpose.”” This comes despite over 600 employees urging CEO Sundar Pichai to reject the agreement, and marks a dramatic reversal from 2018 when Google pulled out of https://x.com/kimmonismus/status/2049081961222955403

Google Signs Classified AI Deal With Pentagon Amid Employee Opposition — The Information https://www.theinformation.com/articles/google-signs-classified-ai-deal-pentagon-amid-employee-opposition

Google shifts Gemini to flexible credit system for usage
Google is replacing fixed subscription tiers with a monthly credit allowance for Gemini, matching the consumption model used by OpenAI and Anthropic. The change would let power users budget more predictably for intensive tasks and give Google a pricing option between its $19.99 and $249.99 tiers. The company is also adding a dedicated image generation section to Gemini, signaling a return of its image editing tools to the core app rather than separate labs.

Google prepares credit system for Gemini and new image tools https://www.testingcatalog.com/google-prepares-credit-system-for-gemini-and-new-image-tools-2/

Google splits AI chip design for training and inference specialization
Google launched two eighth-generation custom chips instead of one unified design—the TPU 8t for training massive models and TPU 8i for fast inference. This marks a strategic shift driven by the rise of AI agents, which require different hardware demands: the 8t delivers nearly 3x prior performance for training, while the 8i cuts inference latency and costs in half through specialized memory and network architecture. The move reflects how AI infrastructure is maturing beyond one-size-fits-all approaches toward purpose-built systems optimized for specific workloads.

Google just broke a decade-long tradition. At Cloud Next 2026, the company unveiled not one, but two new AI chips, the TPU 8t for training and TPU 8i for inference. For the first time ever, Google is splitting its custom silicon into specialized architectures instead of relying https://x.com/kimmonismus/status/2048745304007299230

Two chips for the agentic era https://blog.google/innovation-and-ai/infrastructure-and-cloud/google-cloud/eighth-generation-tpu-agentic-era/

Google commits $10 billion to rival AI startup Anthropic, potentially reaching forty billion.
Google is investing at least $10 billion in Anthropic—a competitor in AI models—with the potential to reach $40 billion if performance targets are met, mirroring a similar $5 billion Amazon deal announced days earlier. Both investments value Anthropic at $350 billion and aim to address severe compute shortages driving recent outages of Anthropic’s Claude AI products. This represents a strategic shift where tech giants fund AI rivals to create a profitable ecosystem where startups become major customers for cloud computing and chips.

*GOOGLE PLANS TO INVEST UP TO $40 BILLION IN ANTHROPIC https://x.com/zerohedge/status/2047704883982180609

Breaking news: Despite offering its own rival Gemini AI models, Google has committed to invest $10bn in Anthropic at its current valuation with a further $30bn to come in the future. https://x.com/FT/status/2047715653553942997

Google Plans to Invest Up to $40 Billion in Anthropic – Bloomberg https://www.bloomberg.com/news/articles/2026-04-24/google-plans-to-invest-up-to-40-billion-in-anthropic?srnd=phx-technology

Google will invest as much as $40 billion in Anthropic – Ars Technica https://arstechnica.com/ai/2026/04/google-will-invest-as-much-as-40-billion-in-anthropic/

Google shifts TPU chip strategy to directly challenge Nvidia’s dominance
Google announced plans to sell its custom AI chips directly to select customers for installation in their own data centers, marking a strategic shift from its previous cloud-rental model. The move, revealed during first-quarter earnings, aims to expand Google’s addressable market as demand grows from AI labs and financial firms, while Amazon simultaneously pushes its own chips with a rumored $50 billion annual revenue run rate, collectively signaling that cloud giants are finally moving beyond Nvidia’s near-monopoly in AI infrastructure hardware.

Google to sell TPU chips to ‘select’ customers in latest shot at Nvidia https://finance.yahoo.com/markets/stocks/article/google-to-sell-tpu-chips-to-select-customers-in-latest-shot-at-nvidia-214900221.html

Hugging Face launches AI agent that automates machine learning research workflows end-to-end.
Hugging Face released ml-intern, an open-source AI agent designed to automate the tedious middle stages of machine learning research—reading papers, tracing citations, building datasets, running experiments, and iterating on results—traditionally handled by human researchers. Early demonstrations show measurable improvements across scientific benchmarks, suggesting the tool could meaningfully compress timelines for model development and free researchers to focus on higher-level strategy rather than execution grunt work.

The @huggingface ML Intern is trending #1 across 1.2M Spaces on the Hub 🔥 What part of my job should we automate next? Link to the demo ⬇️ https://x.com/_lewtun/status/2049021398312468815

.@huggingface unveiled ml-intern – an open-source agent that automates the gritty post-training loop: – reading papers – tracing citations – curating datasets – running experiments – and iterating like a seasoned researcher Early demos show eyebrow-raising gains across science, https://x.com/TheTuringPost/status/2049096050607300765

China blocks Meta’s $2 billion acquisition of AI startup Manus.
China’s state economic planner blocked Meta’s December 2025 purchase of Manus, a Singapore-based AI agents company founded by Chinese engineers, ordering both parties to unwind the deal entirely. The move signals Beijing’s growing control over cross-border AI deals and threatens Meta’s expansion in the fast-moving AI agents sector, while also signaling to Chinese founders that relocating offshore may not shield them from government scrutiny.

China blocks Meta’s $2 billion takeover of AI startup Manus https://www.cnbc.com/2026/04/27/meta-manus-china-blocks-acquisition-ai-startup.html

China blocks Meta’s $2B Manus deal after months-long probe | TechCrunch https://techcrunch.com/2026/04/27/china-vetoes-metas-2b-manus-deal-after-months-long-probe/

Meta commits tens of millions of Graviton chips to power agent AI systems.
Meta is deploying tens of millions of custom AWS Graviton processors to handle the CPU-intensive work that autonomous AI agents require—like reasoning, task planning, and code generation—rather than the GPU-focused training used for traditional AI models. This deal, expanding Meta’s long-standing AWS partnership, reflects a strategic shift in how large-scale AI infrastructure is built, with purpose-built chips optimizing both performance and energy efficiency as agentic AI becomes computationally central to Meta’s operations.

Meta expands Amazon partnership with AWS Graviton chips for AI https://www.aboutamazon.com/news/aws/meta-aws-graviton-ai-partnership

Meta Partners With AWS on Graviton Chips to Power Agentic AI https://about.fb.com/news/2026/04/meta-partners-with-aws-on-graviton-chips-to-power-agentic-ai/

Today we’re announcing an agreement with Amazon Web Services to bring tens of millions of AWS Graviton cores to our compute portfolio. This partnership marks an expansion of our diversified AI infrastructure and will help scale systems behind Meta AI and agentic experiences that https://x.com/AIatMeta/status/2047647617681957207

Meta launches Muse Spark model, signaling shift toward paid AI services.
Meta unveiled Muse Spark, its first major AI model from a rebuilt internal team, marking a strategic pivot from its free, open-source Llama approach toward a paid developer model similar to OpenAI and Google. While early performance benchmarks show the model trails competitors like Anthropic’s Claude, analysts view it as evidence Meta is re-entering the AI race after months of concerns about delays and massive spending ($115–$135 billion planned for 2026), with Wall Street now awaiting clarity on how the company will monetize AI beyond advertising.

Meta Muse Spark has promise, Wall Street wants Zuckerberg AI strategy https://www.cnbc.com/2026/04/28/meta-muse-spark-has-promise-wall-street-wants-zuckerberg-ai-strategy.html

Meta launches AI connectors enabling ad management through third-party platforms.
Meta released integrations that let advertisers control their Meta ads directly from existing AI tools and business software, rather than logging into Meta’s dashboard separately. This matters because it reduces friction in ad workflows for agencies and marketers already using other platforms—similar to how Zapier connects disparate apps. The move signals Meta’s shift toward embedding itself in existing business ecosystems rather than requiring users to come to its own interface.

Introducing Meta Ads AI Connectors: Manage Your Meta Ads From the AI Tools You Already Use | Meta for Business https://www.facebook.com/business/news/meta-ads-ai-connectors

Thinking Machines Lab poaches Meta researchers as Google cloud deal accelerates startup growth.
The AI startup secured a multibillion-dollar infrastructure partnership with Google and access to cutting-edge Nvidia chips, positioning it alongside Anthropic and Meta. Meanwhile, TML has aggressively recruited top Meta talent—including Soumith Chintala, creator of PyTorch, and Piotr Dollár, co-author of Segment Anything—mirroring Meta’s simultaneous hiring from TML’s founding team. The $12 billion valuation and substantial financial upside are drawing seasoned researchers away from Meta, despite its reputation for seven-figure compensation packages.

Meta’s loss is Thinking Machines’ gain | TechCrunch https://techcrunch.com/2026/04/24/metas-loss-is-thinking-machines-gain/

NVIDIA releases compact multimodal AI model handling video, audio, images, and text together.
NVIDIA’s Nemotron 3 Nano Omni is a 30-billion-parameter open model that processes video, audio, images, and text in a single inference loop—a significant step toward practical multimodal AI agents. The model achieves 9× higher throughput than comparable competitors while running on modest hardware (~25GB RAM), making it accessible for enterprises building document intelligence and computer-vision applications. It’s now deployed across major AI platforms including Together AI, Fireworks, OpenRouter, and Fal.

@NVIDIA Nemotron 3 Nano Omni is now on Together AI. Enterprise multimodal AI — video, audio, image, documents & text — optimized for speed and scale. ✅ ~3B active params, 9x higher throughput ✅ Fully managed, zero infra headache ✅ Secure, zero-trust architecture Build https://x.com/togethercompute/status/2049160446708711883

Excited to support @NVIDIA Nemotron 3 Nano Omni, now available on Fireworks. It’s the first open model that handles vision, audio, video, and text in a single inference loop. Built for multimodal sub-agents at scale, with 9× higher throughput than Qwen3 30B. 256K context. Now https://x.com/FireworksAI_HQ/status/2049159136802398546

Introducing @NVIDIA Nemotron 3 Nano Omni. NVIDIA Nemotron 3 Nano Omni is an open multimodal foundation model that unifies audio, images, text, and video into a single context window. It powers subagents for use cases like computer-use agent, document intelligence, and video and https://x.com/baseten/status/2049160818575749300

Introducing NVIDIA Nemotron 3 Nano Omni: Long-Context Multimodal Intelligence for Documents, Audio and Video Agents https://huggingface.co/blog/nvidia/nemotron-3-nano-omni-multimodal-intelligence

Meet Nemotron 3 Nano Omni 👋 Our latest addition to the Nemotron family is the highest efficiency, open multimodal model with leading accuracy. 30B parameters. 256K context length. 🧵👇 https://x.com/NVIDIAAI/status/2049159441870717428

Meet Nemotron 3 Nano Omni 👋 Our latest addition to the Nemotron family is the highest efficiency, open multimodal model with leading accuracy. 30B parameters. 256K context length. 🧵👇 https://x.com/NVIDIAAI/status/2049159441870717428?s=20

NVIDIA Nemotron 3 Nano Omni is now live on fal, available at launch. A single model for multimodal agents: 🔁 text, image, video, audio in one loop 🧠 1 context reasoning across complex workflows ⚡️ ~9× higher throughput with fewer inference hops Built for real-world agent https://x.com/fal/status/2049160999442198632

NVIDIA Nemotron™ 3 Nano Omni is live on OpenRouter. An open 30B-A3B multimodal model for agentic workflows: text, image, video, and audio in → text out, with a 256k context window and efficient MoE architecture for computer use, documents, and AV reasoning. https://x.com/OpenRouter/status/2049164366218772526

NVIDIA releases Nemotron-3-Nano-Omni, a new 30B open multimodal MoE model. Nemotron-3-Nano-Omni-30B-A3B is the strongest omni model for its size and supports audio, video, image and text. Run on ~25GB RAM. GGUF: https://t.co/t4COCqVrLS Guide: https://x.com/UnslothAI/status/2049161390150365344

Figma’s AI plugin automates conversion of text plans into visual boards.
OpenAI’s Codex can now transform written implementation plans directly into interactive FigJam boards through a new Figma plugin, streamlining the handoff between planning and design work. This matters because it removes a manual step that typically requires designers to manually recreate or restructure text-based strategies into visual formats, potentially accelerating product development cycles. The capability demonstrates AI moving beyond code generation into practical workflow automation across design tools.

With the Figma plugin, Codex can now turn implementation plans into visual FigJam boards. https://x.com/OpenAIDevs/status/2049605820351230158

OpenAI commits to five principles guiding AI’s societal role.
In a statement of operating principles, OpenAI pledges to democratize AI access, prevent power concentration among a few companies, empower individual users, pursue universal prosperity through widespread AI infrastructure, build resilience against new risks through collaboration, and remain adaptable as the technology evolves unpredictably. The company argues that these principles—not technical capability alone—will determine whether advanced AI benefits humanity broadly or concentrates power dangerously, framing decisions like massive compute investment and global datacenters as essential to ensuring wide economic participation in an AI-driven future.

Our Principles: Democratization, Empowerment, Universal Prosperity, Resilience, and Adaptability https://x.com/sama/status/2048552677433643427

Our principles | OpenAI https://openai.com/index/our-principles/

OpenAI surpasses AI computing goal one year ahead of schedule.
OpenAI has already exceeded its 2029 target of securing 10 gigawatts of AI computing power—a threefold increase in the past 90 days alone—to support rapidly growing demand for AI services globally. The company frames this infrastructure buildout as essential to democratizing AI benefits and has launched community programs in locations like Abilene, Texas and Wisconsin to create local jobs and manage environmental impacts. This pace of expansion reveals how quickly the bottleneck in AI development has shifted from software capability to physical computing capacity.

Building the compute infrastructure for the Intelligence Age | OpenAI https://openai.com/index/building-the-compute-infrastructure-for-the-intelligence-age/

OpenAI ditches ownership model, pivots to leasing compute capacity instead.
OpenAI has abandoned plans to directly own data centers under its $500 billion Stargate venture with Oracle and SoftBank, instead opting to lease computing power from third parties. The shift reflects the startup’s cash burn concerns—it has missed internal revenue targets and analysts estimate it could run out of money by mid-2027 despite raising $110 billion. The pivot has frustrated partners and raised questions about OpenAI’s reliability, though it mirrors a broader divide: unprofitable AI startups like OpenAI and Anthropic must chase external funding, while established tech giants like Microsoft and Google can self-fund massive infrastructure investments.

OpenAI has effectively abandoned first-party Stargate data centers in favor of more flexible deals — company now prefers to lease compute and says Stargate is an umbrella term | Tom’s Hardware https://www.tomshardware.com/tech-industry/artificial-intelligence/openai-has-effectively-abandoned-first-party-stargate-data-centers-in-favor-of-more-flexible-deals-company-now-prefers-to-lease-compute-and-says-stargate-is-an-umbrella-term

I appreciate you sharing this, but I’m unable to produce a meaningful summary from this material. The text references internal GitHub metrics (“clawsweeper” and “clownfish”) without explaining what these tools are, what problem they solve, or why these numbers matter to readers outside the organization. Without context about whether these represent AI-driven automation, what the business impact is, or how this differs from previous approaches, I cannot write an accurate, informative headline and summary for a business-focused AI newsletter.
Could you provide additional details, such as: What are clawsweeper and clownfish? Are they AI tools? What do these closed issues/PRs represent in practical terms?

Excited that GitHub shows real numbers here again. We been closing over 10k issues and close to 5k PRs this week thanks to clawsweeper and clownfish. Overall since December: 27k issues / 30k PRs closed. https://x.com/steipete/status/2048478136824738181

Claude’s latest update expands AI agents beyond text interfaces significantly.
OpenAI’s newest Claude version now handles voice calls directly with its reasoning engine, adds DeepSeek model support, and improves browser automation with more precise clicking and error recovery. This matters because it removes friction from deploying AI agents across communication platforms—Telegram, Slack, and others—making them more practical for real-world business workflows without workarounds. The shift signals a move toward AI agents that operate across multiple channels and handle more complex interactions than chatbots traditionally have.

OpenClaw 2026.4.24 🦞 ☎️ Voice calls can now reach the full agent 🧠 DeepSeek V4 Flash + Pro join the team 🖱️ Browser automation got coordinate clicks + better recovery 🔧 Telegram, Slack, MCP, sessions, and TTS fixes More reach. Less duct tape. https://x.com/openclaw/status/2048124737918751035

AI agents will soon move faster than humans can observe or control.
Companies are deploying AI systems that can autonomously navigate computers and write code, but these tools currently operate at human speed—slow enough to monitor and correct. The concern is that this window of observability will close rapidly as AI agents become capable of executing tasks at machine speed, potentially before safety oversight mechanisms catch up, fundamentally changing how businesses manage AI systems.

there will this brief era where we can watch our AIs bumble around on the computer clicking things, failing sometimes, taking a ~human amount of time to write code. in the blink of an eye they’ll be manipulating computers far too quickly to monitor https://x.com/tszzl/status/2047766300756488675

AI still needs humans to handle dozens of small, specific gaps.
Current AI systems excel at broad tasks but stumble on countless small details—from simple physical repositioning to nuanced judgment calls—that require human intervention. This fragmented reality reveals why AI integration remains labor-intensive today: the technology handles maybe 80% of a workflow, leaving workers to patch together the remaining scattered pieces rather than automating entire processes end-to-end.

The only way to fully appreciate the jaggedness of the AI frontier is up close. When you use it for a task you know well you find tons of tiny points where AI requires human help. Some are tedious (move a thing) & some profound (is this idea good)? But there are many, for now. https://x.com/emollick/status/2048579531217203375

AI model streams raw pixels directly to screen, bypassing traditional web code.
A team built a prototype that eliminates HTML, layout engines, and conventional code by having an AI model generate pixel-by-pixel what users should see. This approach is distinctive because it replaces the entire infrastructure of web browsers with direct model output, potentially simplifying how interfaces are created but raising questions about accessibility, debugging, and control compared to today’s code-based approach.

Zain Shah on X: “Imagine every pixel on your screen, streamed live directly from a model. No HTML, no layout engine, no code. Just exactly what you want to see. @eddiejiao_obj, @drewocarr and I built a prototype to see how this could actually work, and set out to make it real. We’re calling it https://t.co/C4BEi1lse8” / X https://x.com/zan2434/status/2046982383430496444

Taylor Swift trademarks her voice and image to fight AI misuse.
Swift filed three trademark applications—two protecting distinctive phrases of her voice and one covering a specific visual of her in performance attire—following a legal strategy pioneered by Matthew McConaughey. The move matters because trademarks could provide stronger federal legal grounds to combat unauthorized AI-generated content than existing state-level right-of-publicity laws, giving entertainers a new tool to prevent their likenesses from being stolen by AI systems. Swift’s action comes after her image was used without consent in AI fakes by Meta and in deepfake pornography, underscoring why this protection has become urgent for high-profile figures.

Taylor Swift Files to Trademark Voice and Likeness to Protect Against AI Misuse https://variety.com/2026/music/news/taylor-swift-trademark-voice-likeness-ai-misuse-1236731401/

Figure achieves major scaling milestone, producing humanoid robot every hour.
Figure AI has ramped production of its humanoid robots from one per day to one per hour over four months, hitting 55 units this week—a concrete demonstration that the company can move beyond prototypes to actual manufacturing at meaningful volumes. This matters because scaling industrial robotics from lab to factory floor has historically been the hardest part; many promising robot startups stalled when facing real production constraints. The 24x acceleration in just 120 days suggests either significant process improvements or that Figure solved critical manufacturing bottlenecks that typically plague hardware startups.

In the last 120 days, Figure scaled manufacturing 24x – from 1 robot/day to 1 robot/hour We will manufacture 55 humanoid robots this week https://x.com/adcock_brett/status/2049514372264055116

I appreciate the enthusiasm, but I need substantive material to summarize—this message doesn’t contain reporting or verifiable information about what happened, why it matters, or the evidence behind the claim.
Could you provide: – A news article, press release, or detailed description of the Unitree G1 development? – Context on what makes this capability significant (e.g., new engineering milestone, commercial availability, comparison to competitors)? – Any concrete details or sources? Once you share that, I’ll produce the two-line summary in the format you’ve requested.

This is incredibly cool, Unitree G1 on inline skates! https://x.com/TheHumanoidHub/status/2047345074011586655

Xiaomi releases open-source MiMo-V2.5-Pro agent model for complex coding tasks.
Xiaomi open-sourced MiMo-V2.5-Pro, a 1 trillion-parameter AI model designed to autonomously handle multi-step engineering and coding projects that typically require weeks of human expertise. The model demonstrated its capability by building a Rust compiler from scratch in 4.3 hours and an 8,000-line video editor in 11.5 hours, while using 40–60% fewer tokens than comparable frontier models like Claude Opus 4.6. Developers gain immediate access via vLLM support and a 100 trillion free token grant for builders.

🎉 Day-0 vLLM support for the MiMo-V2.5 series! Congrats to @XiaomiMiMo on the open-source release of the MiMo-V2.5 and MiMo-V2.5-Pro. Highlights from the flagship MiMo-V2.5-Pro, an agent-oriented model focused on long-horizon tool use and frontier coding: – Long-horizon task https://x.com/vllm_project/status/2048825703244972375

Just dropped two open-source models: MiMo-V2.5-Pro (Code Agent, 1T total) and MiMo-V2.5 (Multimodal Agent, 310B total). Oh and one more thing — we’re giving devs & creators 100T tokens on us. Go build something cool 🛠️ 🎁 100T Free Token Grant for Builders https://x.com/_LuoFuli/status/2048851054662762618

MiMo-V2.5-Pro | Xiaomi https://mimo.xiaomi.com/mimo-v2-5-pro

Xiaomi MiMo-V2.5 is now officially open-sourced! MIT License, supporting commercial deployment, continued training, and fine-tuning – no additional authorization required. Two models, both supporting a 1M-token context window : • MiMo-V2.5-Pro: built for complex agent and https://x.com/XiaomiMiMo/status/2048821516079661561

Xiaomi MiMo-V2.5 Series: Pushing Open-Source Agents Forward 🔸 MiMo-V2.5-Pro, our strongest model yet. A major leap from MiMo-V2-Pro in general agentic capabilities, complex software engineering, and long-horizon tasks, now matching frontier models like Claude Opus 4.6 and https://x.com/XiaomiMiMo/status/2046988157888209365?s=20

Leave a Reply

Trending

Discover more from Ethan B. Holland

Subscribe now to keep reading and get access to the full archive.

Continue reading