Image created with gemini-3.1-flash-image-preview with claude-sonnet-4-5. Image prompt: Using the provided reference image of the Mercedes hood ornament, preserve the exact midnight navy car hood, chrome pedestal base, shallow depth-of-field sky background, dramatic upward camera angle, and automotive advertisement lighting. Replace only the Mercedes star with a vertical strip of 35mm film standing upright on the pedestal, rendered in the same polished chrome metal with visible sprocket holes and film frames, photorealistic and proportional to a luxury car hood ornament. Add bold white sans-serif text reading VIDEO across the upper portion of the image as a clean headline.
ByteDance reportedly pauses global launch of its Seedance 2.0 video generator | TechCrunch https://techcrunch.com/2026/03/15/bytedance-reportedly-pauses-global-launch-of-its-seedance-2-0-video-generator/
ByteDance Suspends Launch of Video AI Model After Copyright Disputes With Hollywood — The Information https://www.theinformation.com/articles/bytedance-suspends-launch-video-ai-model-copyright-disputes-hollywood
A breakthrough in real-time video generation. As a research preview developed with @NVIDIA and shared at @NVIDIAGTC this week, we trained a new real-time video model running on Vera Rubin. HD videos generate instantly, with time-to-first-frame under 100ms. Unlocking an entirely
https://x.com/runwayml/status/2034284298769985914#m
NVIDIA GTC 2026 Keynote: Everything That Happened in 12 Minutes – YouTube https://www.youtube.com/watch?v=X2i_8O75_Os
Schibsted open sources AI tool that turns news articles into videos | Schibsted https://schibsted.com/news/schibsted-open-sources-ai-tool-that-turns-news-articles-into-videos/
DoorDash’s New Paid Tasks Turn Couriers Into AI and Robot Trainers – Bloomberg https://www.bloomberg.com/news/articles/2026-03-19/doordash-s-new-paid-tasks-turn-couriers-into-ai-and-robot-trainers
Someone used Suno AI to generate a Japanese metal band called Neon Oni. Fake member bios, AI-generated music videos, “”Based in Tokyo”” on Spotify. 80,000+ monthly listeners. Fans had it in their Spotify Wrapped top 5. Merch was selling. Then, community sleuths exposed it. Traced
https://x.com/TheRundownAI/status/2033568236227244451?s=20
Probably the most current look at Palantir’s maven smart system software. Here’s the DoW’s Chief AI officer showing how it works:
https://x.com/bilawalsidhu/status/2032432668105712093
LiTo: Joint Geometry and Appearance Modeling for Image-to-3D Generation TL;DR: Generates high-fidelity 3D objects from a single image by jointly modeling geometry + view-dependent appearance (lighting, reflections) in a unified latent space
https://x.com/Almorgand/status/2033987312451731904
DVD: Dynamic Video Depth”” TL;DR: Recovers temporally consistent depth from monocular videos using diffusion priors + geometric constraints, handling dynamic scenes and motion robustly.
https://x.com/Almorgand/status/2034349445601538057
NVIDIA thanks all its partners: the message? There is no way around NVIDIA. NVIDIA is the center of the revolution.
https://x.com/kimmonismus/status/2033615181415387610
Straight from NVIDIA GTC: Jensen Huang just unveiled a new vision for AI infrastructure For the first time, Rubin GPUs+Groq LPUs are paired: > 35× higher inference throughput > 10× more revenue from trillion-parameter models Architecture & why it’s needed
https://x.com/TheTuringPost/status/2033700480975520097
Thank you Jensen and NVIDIA! She’s a real beauty! I was told I’d be getting a secret gift, with a hint that it requires 20 amps. (So I knew it had to be good). She’ll make for a beautiful, spacious home for my Dobby the House Elf claw, among lots of other tinkering, thank you!!
https://x.com/karpathy/status/2034321875506196585
Regardless of which text-to-video model is best: prompt-to-output in under 100ms is incredible. And many people don’t grasp its significance. Time is money. Instead of waiting half a minute or more for the output each time, only to make changes again, direct prompt-to-output is
https://x.com/kimmonismus/status/2034659158843072893
AI assistants are mapping our inner world. Spatial intelligence is mapping the outer one. Layer them together and this is what you see:
https://x.com/bilawalsidhu/status/2033225076116529543
Big moment. You can cross the uncanny valley in video games by using real-time video-to-video AI. You get the best of coherence & control from classical 3d engines, then use generative AI to take it all the way.
https://x.com/bilawalsidhu/status/2033627865300816326
One of the most prescient scenes in movie history. Radio frequency (RF) is the next big modality for spatial intelligence.
https://x.com/bilawalsidhu/status/2033009623817416955
When your 3d scan is like replaying a dream
https://x.com/bilawalsidhu/status/2033257427773095936
At this nerdiest of all nerdy sessions 💞, Jeff Dean said he doesn’t think we’re running out of data. “I think there’s still an enormous amount of data in the world that we haven’t really used yet for training these models. We train on some video data, for example, but there’s a
https://x.com/TheTuringPost/status/2034411360302567803
Learning from robot data? Standard. Direct Video-Action Models (DVA) is different: treat robot control as video generation, then translate the generated video into actions. Built by @rhoda_ai_, the system pre-trains causal video models from scratch and can run complex
https://x.com/IlirAliu_/status/2032742738853048413
Video generation might be a much better backbone for robot learning than image-text models. DiT4DiT couples a video Diffusion Transformer with an action Diffusion Transformer, letting robot policies learn directly from spatiotemporal video dynamics instead of static visual
https://x.com/IlirAliu_/status/2032380216962691114





Leave a Reply