Image created with gemini-3.1-flash-image-preview with claude-sonnet-4-5. Image prompt: Using the provided reference image, preserve the exact compositional structure with subject in left third and smoky right two-thirds, the deep blue-purple cinematic lighting, atmospheric haze, and emotional gravity, but replace the central figure with a vintage Polaroid instant camera in dramatic close-crop angled slightly downward, lens reflecting blue-purple light, fine iridescent glitter scattered across its black body, a just-ejected overexposed photograph hanging from the film slot, and replace the title text with ‘images’ in thin lowercase white Helvetica Neue Light on the misty right side.
People are asking what’s the difference between Falcon Perception and SAM3, so here’s my opinion: SAM3:
https://t.co/KVRbuHm8H1 Falcon Perception:
https://t.co/QDgMlOBvDH First, sam3 does “”promptable concept segmentation””: simple noun phrases (like “”yellow bus””, “”red apple””) +
https://x.com/dahou_yasser/status/2041474094252933195
Today we’re releasing WildDet3D–an open model for monocular 3D object detection in the wild. It works with text, clicks, or 2D boxes, and on zero-shot evals it nearly doubles the best prior scores. 🧵
https://x.com/allen_ai/status/2041545111151022094
kays on X: “I noticed there wasn’t anything like this out there, so I wrote a tiny visual blog for those wanting to introduce themselves to Dynamic Gaussian Splatting and their current methods 🖼️ Feel free to check out, these are some of the visuals taken from it https://t.co/6W2qx2yI1K” / X
https://x.com/pabloadaw/status/2041650303804555278
We’re excited to be rolling out two model updates today! Marble 1.1: Improves lighting and contrast, with a major reduction in visual artifacts. Marble 1.1-Plus: Our new model built for scale. Create larger, more complex environments than ever before.
https://x.com/theworldlabs/status/2041554646561677701
Generate 3D models and interactive charts with the Gemini app
https://blog.google/innovation-and-ai/products/gemini-app/3d-models-charts/
Breaking: @AIatMeta just released Muse Spark — now live across @ScaleAILabs leaderboards. Here’s how it stacks up: Tied for 🥇on SWE-Bench Pro Tied for 🥇on HLE Tied for 🥇on MCP Atlas Tied for 🥇on PR Bench – Legal Tied for 🥈on SWE Atlas Test Writing 🥈on PR Bench – Finance
https://x.com/scale_AI/status/2041934840879358223
Introducing Muse Spark, the first in the Muse family of models developed by Meta Superintelligence Labs. Muse Spark is a natively multimodal reasoning model with support for tool-use, visual chain of thought, and multi-agent orchestration. Muse Spark is available today at
https://x.com/AIatMeta/status/2041910285653737975
NEW: Meta announces Muse Spark. All you need to know: * It’s their new multi-modal reasoning model. * Strong at multi-agent orchestration and multi-modal reasoning. * Contemplating mode orchestrates multiple agents that reason in parallel. Helps to compete with models such
https://x.com/omarsar0/status/2041919769536770247
To spend more test-time reasoning without drastically increasing latency, we can scale the number of parallel agents that collaborate to solve hard problems. While standard test-time scaling has a single agent think for longer, scaling Muse Spark with multi-agent thinking enables
https://x.com/AIatMeta/status/2041926297216282639
Meta is back! Muse Spark scores 52 on the Artificial Analysis Intelligence Index, behind only Gemini 3.1 Pro, GPT-5.4, and Claude Opus 4.6. Muse Spark is the first new release since Llama 4 in April 2025 and also Meta’s first release that is not open weights Muse Spark is a new
https://x.com/ArtificialAnlys/status/2041913043379220801
try muse spark via the Meta AI app or
https://t.co/DipeeIuXm2! check out this simulation i made:
https://x.com/alexandr_wang/status/2041953243895623913
1/ today we’re releasing muse spark, the first model from MSL. nine months ago we rebuilt our ai stack from scratch. new infrastructure, new architecture, new data pipelines. muse spark is the result of that work, and now it powers meta ai. 🧵
https://x.com/alexandr_wang/status/2041909376508985381
The new model from Meta, Muse Spark, is pretty good at converting images to code!
https://x.com/skirano/status/2041920891072700631
Excited to share what we’ve been building at Meta Superintelligence Labs! We just released Muse Spark, our first AI model. It’s a natively multimodal reasoning model and the first step on our path to personal superintelligence. We’ve overhauled our entire stack to support
https://x.com/shengjia_zhao/status/2041909050728931581
Introducing Muse Spark: Scaling Towards Personal Superintelligence
https://ai.meta.com/blog/introducing-muse-spark-msl/
Meta is back in the game! It’s been fun to test out Muse Spark. Beyond benchmarks, it’s actually a good day to day model… surprisingly good at technical problems and making arcade games. Never bet against @alexandr_wang @natfriedman @danielgross
https://x.com/matthuang/status/2041911766586945770
Meta just released a frontier model, Muse Spark- it takes the #3 spot on our Vals Index.
https://x.com/ValsAI/status/2041922037745381389
try muse spark yourself! download the Meta AI app or go to
https://x.com/alexandr_wang/status/2042024651610861657
We had pre-release access to Meta’s new Muse Spark model and evaluated it on FrontierMath. It scored 39% on Tiers 1-3 and 15% on Tier 4. This is competitive with several recent frontier models, though behind GPT-5.4.
https://x.com/EpochAIResearch/status/2041947954202988757
To build personal superintelligence, our model’s capabilities should scale predictably and efficiently. Below, we share how we study and track Muse Spark’s scaling properties along three axes: pretraining, reinforcement learning, and test-time reasoning. 🧵👇 Let’s start with
https://x.com/AIatMeta/status/2041926291142930899
I showed you SAM 3 all week. This is a 0.6B model that outperforms it. Falcon Perception. Type “”detect the plane”” and it segments every plane in the frame. Pixel-accurate masks from natural language. Fighter jets. Fire. Crowds. All on a MacBook via MLX. No cloud.
https://x.com/MaziyarPanahi/status/2040776481673281936
OpenAI tests next-gen Image V2 model on ChatGPT and LM Arena
https://www.testingcatalog.com/openai-tests-next-gen-image-v2-model-on-chatgpt-and-lm-arena/
PoseDreamer: Scalable and Photorealistic Human Data Generation Pipeline with Diffusion Models”” TL;DR: diffusion pipeline for scalable generation of photorealistic human data with 3D annotations
https://x.com/Almorgand/status/2040096997470843366
StereoVGGT: A Training-Free Visual Geometry Transformer for Stereo Vision”” TL;DR: adapts a pretrained 3D-aware transformer to stereo vision with a training-free pipeline, achieving SOTA performance on KITTI
https://x.com/Almorgand/status/2041569246883332385
Subtle hand held camera shake will convince people a 3d game render is real life footage. No generative ai required.
https://x.com/bilawalsidhu/status/2041643400433201384
PoseDreamer: Scalable Photorealistic Human Data Generation with Diffusion Models
https://prosperolo.github.io/posedreamer/
We always need more visuals! Checkout this on for dynamic gaussian splatting
https://x.com/Almorgand/status/2041773431524302968
Seems like a good model from Meta that is still trailing the current series of releases. The most important thing to note is that it is not open weights. That was the main reason that Meta’s models were so important. Without that, it is a lot harder to predict the value of Spark
https://x.com/emollick/status/2041924282964394085
try for yourself!
https://t.co/DipeeIuXm2 or download Meta AI app
https://x.com/alexandr_wang/status/2041985846950424760
Our first model from MSL, Muse Spark, is now available on
https://t.co/qBMQ6BPVgP! This is an efficient all-rounder model. It supports fast responses, deeper thinking, visual chain of thought, a higher inference “Contemplating” mode. Plus, it’s natively multimodal. 1/
https://x.com/jack_w_rae/status/2041925332631183421
1/ It’s been so fun working with @shengjia_zhao, @alexandr_wang and the team to build muse spark from scratch. It is early and has rough edges, but excited to continue our research velocity. I especially love that we’re doubling down on the fundamental science. We’re focused on
https://x.com/ananyaku/status/2041913147842556390
1/ Muse Spark is live, and alongside it, our new Advanced AI Scaling Framework which details how we evaluate and prepare for advanced AI. We tested across bio, chem, cyber, and loss of control risks before and after mitigations. Muse Spark achieves a 98% bioweapons refusal rate
https://x.com/summeryue0/status/2041956901769113948
Check out Muse Spark, our first milestone in the quest for personal superintelligence! Scaling this with the team has been a total blast. Give it a spin and let us know what you think! 🥑
https://x.com/ren_hongyu/status/2041922484040298796
e
try muse spark on
https://x.com/alexandr_wang/status/2041956770864885870
The team really killed it on photorealism with this model. While testing we found out ice is a great barometer for this – try this glacier prompt. A glacier wall towering like a cathedral interior, deep blue ice with light refracting through layers, tiny human figure at base for
https://x.com/mustafasuleyman/status/2041210297491574861
Introducing Quality mode on Grok Imagine – powered by our most advanced image generation model. Quality mode gives you enhanced details, stronger text rendering, and higher levels of creative control. Now available on web and mobile. Try it at
https://x.com/xai/status/2040150187373670570
We’ve added a new pseudonymous video model to our Text to Video and Image to Video Arenas.’HappyHorse-1.0′ is currently landing in the #1 spot for Text and Image to Video (No Audio) and the #2 spot for Text and Image to Video (With Audio). Further details coming soon. Example
https://x.com/ArtificialAnlys/status/2041591989083500933





Leave a Reply