Image created with gemini-3.1-flash-image-preview with claude-sonnet-4-5. Image prompt: Animation cel style illustration of a muscular blue genie emerging from a golden oil lamp, magical teal wisps flowing from his hands onto a curved 35mm film strip, individual frames transforming from blank to colorful animated scenes, Disney-quality hand-drawn aesthetic with clean lines, jewel tone colors, warm cinematic lighting, simple gradient background, horizontal composition with space for title text across top third.
I got early access to Project Genie from @GoogleDeepMind ✨ It’s unlike any realtime world model I’ve tried – you generate a scene from text or a photo, and then design the character who gets to explore it. I tested dozens of prompts. Here are the standout features 👇”” https://x.com/venturetwins/status/2016919922727850333
HOLY FUCK Genie 3 is the craziest thing I’ve tried in a long time Just… wow. Watch this.”” https://x.com/mattshumer_/status/2017058981286396001
Project Genie is an impressive demonstration of what world models can do. But there’s a difference between seeing the future and being able to build with it today. This is what running locally looks like”” https://x.com/overworld_ai/status/2017298592919392717
Here’s how it works: 🔵 Design your world and character using text and visual prompts. 🔵 Nano Banana Pro makes an image preview that you can adjust. 🔵 Our Genie 3 world model generates the environment in real-time as you move through. 🔵 Remix existing worlds or discover new”” https://x.com/GoogleDeepMind/status/2016919762924949631
Project Genie is a prototype web app powered by Genie 3, Nano Banana Pro + Gemini that lets you create your own interactive worlds. I’ve been playing around with it a bit and it’s…out of this world:) Rolling out now for US Ultra subscribers.”” https://x.com/sundarpichai/status/2016979481832067264
5/ Building responsibly 🛡️ Building AI responsibly is core to our mission. As an experimental @GoogleLabs prototype, Project Genie is still in development. This means you might encounter 60-second generation limits, control latency, or physics that don’t always perfectly adhere”” https://x.com/Google/status/2016972686208225578
Project Genie: AI world model now available for Ultra users in U.S. https://blog.google/innovation-and-ai/models-and-research/google-deepmind/project-genie/
Thrilled to launch Project Genie, an experimental prototype of the world’s most advanced world model. Create entire playable worlds to explore in real-time just from a simple text prompt – kind of mindblowing really! Available to Ultra subs in the US for now – have fun exploring!”” https://x.com/demishassabis/status/2016925155277361423
Introducing Project Genie: An experimental research prototype powered by Genie 3, our world model, that lets you prompt an interactive world into existence — and then step inside 🌎”” https://x.com/Google/status/2016926928478089623
Project Genie is rolling out for AI Ultra members in the USA. It’s an experimental tool that allows you to create and explore infinite virtual worlds, and I’ve never seen anything like this. It’s still early, but it’s already unreal. Nano Banana Pro + Project Genie = My low-poly”” https://x.com/joshwoodward/status/2016921839038255210
Step inside Project Genie: our experimental research prototype that lets you create, edit, and explore virtual worlds. 🌎”” https://x.com/GoogleDeepMind/status/2016919756440240479
Project Genie is rolling out to @Google AI Ultra subscribers in the U.S. (18+) With this prototype, we want to learn more about immersive user experiences to advance our research and help us better understand the future of world models. See the details → https://x.com/GoogleDeepMind/status/2016919765713826171
I’ve written 250k+ lines of game engine code. Here’s why Genie 3 isn’t what people think it is: World models are something genuinely new. A third category of media we don’t have a name for yet. Near-term they’re too slow and expensive for consumers. But for training robots?”” https://x.com/jsnnsa/status/2017276112561422786
xAI’s Grok Imagine takes the #1 spot in both Text to Video and Image to Video in the Artificial Analysis Video Arena, surpassing Runway Gen-4.5, Kling 2.5 Turbo, and Veo 3.1! Grok Imagine is the latest video model from @xAI, and joins an increasing roster of models such as”” https://x.com/ArtificialAnlys/status/2016749756081721561
Vidu Q3 Pro ranks #2 in Text to Video in the Artificial Analysis Video Arena, surpassing Runway Gen-4.5 and Kling 2.5 Turbo while trailing only xAI’s Grok Imagine! Vidu Q3 Pro is the latest release from @ViduAI_official, representing a significant upgrade from their Vidu Q2″” https://x.com/ArtificialAnlys/status/2017225053008719916
🚨BREAKING: @xAI’s first model in Video Arena debuts in the top 3! Grok-Imagine-Video ranks #3 on the Image-to-Video Arena and #4 on the Text-to-Video Arena. It is close to the top-ranked @GoogleDeepMind Veo 3.1 and @OpenAI Sora 2 Pro models. Grok-Imagine-Video offers: -“” https://x.com/arena/status/2016748418635616440
@xai Try New Grok Imagine here! Text to Image https://t.co/OeJMwL9hoH Image Editing https://t.co/Q7lojX41I1 Text to Video https://t.co/fAzEJABTYn Image to Video https://t.co/zTdoJQjkqk Video Editing”” https://x.com/fal/status/2016746473887609118
World Models | Ankit Maloo https://ankitmaloo.com/world-models/
🚨 Hunyuan 3D 3.1 Pro and Rapid is here on fal! 🎯 Pro: High-fidelity Image-to-3D and Text-to-3D generation ⚡ Rapid: Speed-optimized 3D generation ✨ Smart Topology and Part generation for advanced 3D workflows”” https://x.com/fal/status/2016877742298411089
D4RT: Unified, Fast 4D Scene Reconstruction & Tracking — Google DeepMind
https://deepmind.google/blog/d4rt-teaching-ai-to-see-the-world-in-four-dimensions/
Crazy results but the speed is what makes this incredible!”” https://x.com/Almorgand/status/2014615608545915168
Our short film Dear Upstairs Neighbors is previewing at @sundancefest. 🎬 It’s a story about noisy neighbors, but behind the scenes, it’s about solving a huge challenge in generative AI: control. Developed by Pixar alumni, an Academy Award winner, researchers, and engineers,”” https://x.com/GoogleDeepMind/status/2015847703276577251
WildRayZer: Self‑supervised Large View Synthesis in Dynamic Environments”” TL;DR: self‑supervised NVS model that disentangles motion from static structure to render clean static novel views from dynamic video without 3D supervision.”” https://x.com/Almorgand/status/2014754835740958788
Motion 3-to-4: 3D Motion Reconstruction for 4D Synthesis”” TL;DR: feed-forward 4D mesh synthesis from a single monocular video (+ optional reference mesh) by predicting per-frame vertex trajectories for temporally coherent motion.”” https://x.com/Almorgand/status/2014391936178643447
GR3EN: Generative Relighting for 3D Environments”” TL;DR: generative relighting method that distills video-to-video relighting diffusion outputs into 3D scene relighting, enabling controllable lighting changes in large room-scale reconstructions.”” https://x.com/Almorgand/status/2016202951908274228
damn. spatial intelligence can be FAST asf! “”D4RT can continuously understand what’s moving while running 18x-300x faster than previous methods – processing a 1-minute video in roughly 5 seconds on a single TPU chip.”””” https://x.com/bilawalsidhu/status/2014490782506356998
Here’s how to turn the world around you into a story you can tell with Gen-4.5 Image to Video.”” https://x.com/runwayml/status/2017238025982427316
Runway Gen 4.5 has two new features. Motion Sketch and Character Swap are now in-built apps for the tool. Sketch camera motion with annotations on a start frame to control the movement. Swap a character with two images before the video step. Here’s how!”” https://x.com/jerrod_lew/status/2016816309762486423
Take inspiration from your world. Then turn it into a story you want to tell. With Gen-4.5 Image to Video. Simply take a photo, load it into Runway, then ask for what you want. A Day at the Museum. Generated with AI. Made by Áron. Full how to video coming soon. Get started at”” https://x.com/runwayml/status/2016882344427147275
This is happening live. It’s a realtime AI video model. We are so cooked.”” https://x.com/bilawalsidhu/status/2015993354576634235
World models are going to take over in 2026. While video generation models are impressive dreamers, they aren’t world simulators. Generative video models can produce stunning clips, but they hallucinate pixel transitions based on statistical correlations. This leads to”” https://x.com/dair_ai/status/2016881546909929775
Human3R: Everyone Everywhere All at Once”” TL;DR: unified feed-forward 4D reconstruction from monocular video joint multi-person SMPL-X, scene geometry, and camera trajectories in real time (~15 FPS).”” https://x.com/Almorgand/status/2016546477569429544
Whenever I see demos about gaming world models, this is what I expect. Anything else is video generation not gaming”” https://x.com/sethkarten/status/2017322251385745570
Thrilled to share our new Grok Imagine release 🚀 It is the highest quality, fastest, and most cost-effective video generation model yet. Comes with 720P, video editing and better audio! We listened closely to your feedback and moved fast. Just six months ago, we had almost”” https://x.com/EthanHe_42/status/2016749123198673099
Grok Imagine is also #1 in the Artificial Analysis Image to Video Leaderboard!”” https://x.com/ArtificialAnlys/status/2016749790907027726
LingBot-World from Ant Group An open-source world simulator from video generation with real-time interactivity. Maintains high fidelity across diverse environments with minute-level consistency and <1s latency at 16 FPS.”” https://x.com/HuggingPapers/status/2016787043028746284
fal is proud to partner with @xai as Grok Imagine’s day-0 platform partner xAI’s latest image & video gen + editing model ✨ Stunning photorealistic images/videos from text ⚡ Lightning-fast generation 🎥 Dynamic animations with precise control 🎨 Edit elements, styles & more”” https://x.com/fal/status/2016746472931283366
.@xai’s Groks new video generation model is so freaking good. And even more important: price/performance ratio is next level.”” https://x.com/kimmonismus/status/2017252078272553396
LingBot-World is unveiled as an open-source, real-time interactive world model built on Alibaba’s Wan2.2, capable of generating. But heres the catch: nearly 10 minutes of stable, continuous generation – even after the camera looks away for 60 seconds, objects remain intact when”” https://x.com/kimmonismus/status/2016896151610442192
Hands-on way to robotics – Python & simulation.👉 This free course might be what you need [Github ⬇️] Set of tutorials to practice robotics concepts with Webots and Python. Robotics Simulation Labs was created by Felipe N. Martins! An open resource packed with practical”” https://x.com/IlirAliu_/status/2014260753692230131
A generative world for general-purpose robotics & embodied AI learning. Genesis is a physics platform designed for general-purpose Robotics/Embodied AI/Physical AI applications. 📍GitHub: https://t.co/1WkYOD8Djm —– Weekly robotics and AI insights. Subscribe free:”” https://x.com/IlirAliu_/status/2015710305368605022
Foundation models are enough to solve robotics! Unfortunately, this is not true. We keep hearing that Vision-Language-Action (VLA) models struggle because of the gap between static training and the dynamic real world. A German startup (@SereactAI) just released a solution that”” https://x.com/IlirAliu_/status/2016228327103574326
High-speed food packaging only makes sense if automation actually changes the economics. At Anı Bisküvi A.Ş., a robotic box-filling system from Robentex now runs two lines at a combined 800 products per minute. They moved to a tray-and-lid concept instead of classic display”” https://x.com/IlirAliu_/status/2015863130341941749
Today we’re introducing Helix 02 Dancing robots are trivial, the hard part is intelligent control This is our most powerful model to date – able to work across complex tasks & long time horizons”” https://x.com/adcock_brett/status/2016207851891667395
Try Grok Imagine now:”” https://x.com/chaitu/status/2017297699973042412
Grok Imagine API | xAI https://x.ai/news/grok-imagine-api
Grok Imagine only gets better from here”” https://x.com/elonmusk/status/2016768088855769236
Gen-4.5 is pretty much an animation engine at this point”” https://x.com/c_valenzuelab/status/2016721443430510847
Ray3.14 | Luma AI https://lumalabs.ai/blog/news/ray3_14
Video Arena Is Live on Web https://arena.ai/blog/video-arena/
‘Dear Upstairs Neighbors’: Animated film made with Google AI https://blog.google/innovation-and-ai/models-and-research/google-deepmind/dear-upstairs-neighbors/
Realtime | Krea https://www.krea.ai/realtime
We’re helping AI to see the 3D world in motion as humans do. 🌐 Enter D4RT: a unified model that turns video into 4D representations faster than previous methods – enabling it to understand space and time. This is how it works 🧵”” https://x.com/GoogleDeepMind/status/2014352808426807527





Leave a Reply