Image created with gemini-2.5-flash-image with claude-sonnet-4-5. Image prompt: Elegant holiday wrapping paper pattern featuring ornate film strip frames arranged in damask-style repeat, each frame containing evolving imagery from silent film silhouettes to AI-generated scenes, connected by Victorian filigree and motion blur flourishes, Art Deco ‘VIDEO’ marquee typography integrated into pattern, rich burgundy and antique gold with subtle embossed texture, sophisticated gift wrap design in the style of Liberty of London textiles.
The dawn of a world simulator https://odyssey.ml/the-dawn-of-a-world-simulator
Visual sim2real: zero-shot deploy to the real world, with zero real data. Trained entirely in Isaac Lab. https://x.com/DrJimFan/status/2003879976173818298
OpenAI built the Sora Android app (which hit #1 app in the world) in just 18 days with the help of Codex https://x.com/lennysan/status/2001074732293300301
Adobe and Runway Partner to Deliver the Next Generation of AI Video for Creators, Studios and Brands https://news.adobe.com/news/2025/12/adobe-and-runway-partner
腾讯混元 https://hunyuan.tencent.com/motion?tabIndex=0
Soul https://zhangzjn.github.io/projects/Soul/
STARCaster: Spatio-Temporal AutoRegressive Video Diffusion for Identity- and View-Aware Talking Portraits https://foivospar.github.io/STARCaster/
Animate Any Character in Any World”” TL;DR: users’ provided 3DGS scene along with a 3D or multi-view character -> enabling interactive control of the character’s behaviors and active exploration of the environment through natural language commands https://x.com/Almorgand/status/2003518454280687885
FlexAvatar: Learning Complete 3D Head Avatars with Partial Supervision”” TL;DR: transformer-based 3D portrait animation model with learnable data source tokens, so-called bias sinks, which enables unified training across monocular and multi-view datasets. https://x.com/Almorgand/status/2003153695765336468
PAD3R: Pose-Aware Dynamic 3D Reconstruction from Casual Videos”” TL;DR: canonical frame selection;image to 3D (static 3D Gaussian);set of randomly sampled camera poses to fine-tune a lightweight image2pose estimator; camera pose estimator to optimize a deformable 3d object model https://x.com/Almorgand/status/2001695549259747415
There’s something so magical about turning 2d videos into 4d reconstructions. Every video becomes a spatio-temporal portal back in time – one that you can revisit from any angle. Research like d4rt is turning science fiction into reality; and it’s getting fast enough to run in https://x.com/bilawalsidhu/status/2003698903838003685
Virtually Being : Customizing Camera-Controllable Video Diffusion Models with MultiView Performance Captures TLDR: multiview character consistency; 3D camera control in video diffusion models; character trained via 4DGS,lighting variability obtained with a video relighting model https://x.com/Almorgand/status/2002069630622507504
Researchers proposed Sample-Efficient Modality Integration (SEMI), which plugs any pretrained encoder (image, audio, video, sensors, graphs) into an LLM using one projector plus LoRA adapters generated from a handful of paired examples. Trained on data-rich domains, SEMI https://x.com/DeepLearningAI/status/2003593131132916204
Vibe coding, but for robotics. Fully generated. The setup: • Designed with Nano Banana Pro • Built using Gemini 3 • Generated from high-level intent, not low-level code This robot arm simulation can stack cubes and build walls. No traditional programming. No hand-written https://x.com/IlirAliu_/status/2001593025881944387
AI Slop Report: The Global Rise of Low-Quality AI Videos https://www.kapwing.com/blog/ai-slop-report-the-global-rise-of-low-quality-ai-videos/





Leave a Reply