Image created with gemini-3.1-flash-image-preview with claude-opus-4.7. Image prompt: High-end product photograph of a tall parfait glass filled with two mirrored swirls of soft-serve ice cream, one tinted cool cyan and one warm magenta like an anaglyph 3D image, topped with two round waffle cookies positioned like stereoscope lenses and crossed red and cyan spoons in front, a crisp white paper band around the glass printed with bold ‘AR/VR’ lettering and a small ’75 — Milford, DE’ stamp, soft directional studio light, shallow depth of field, glossy macro detail, creamy nostalgic Dairy Queen palette, landscape composition.
Every pixel generated, not rendered” goes far beyond gaming. The model is the render loop and the layout engine. The DOM dissolves – every pixel semantically addressable, every region interactive by default. That is kind of nuts!
https://x.com/bilawalsidhu/status/2047148638753681709
Vista4d – capture something in 2d once; reframe camera moves infinitely in post production. Impressive research by Netflix.
https://x.com/bilawalsidhu/status/2048568784076648553
360 drones are epic for capturing immaculate 3d gaussian splats. Wait till the virtual camera flies through the tree canopy down to ground level – bloody magnificent !
https://x.com/bilawalsidhu/status/2049345616849907885
Okay fine I’m buying a 360 drone – these things look like amazing sensor collection contraptions for making immaculate 3d gaussian splats
https://x.com/bilawalsidhu/status/2047907181253562568
one 2d photo –> 3d gaussian splat quick test with Echo-2 by SpAItial — these 3d scene generation models are getting better! already at a sufficient quality to serve as a virtual set / backdrop in your 3d tool of choice
https://x.com/bilawalsidhu/status/2049615166434586773
The same satellite tech can track a convoy through dense cloud cover at night, flag a sinkhole weeks before it opens, catch a village built on a collapsing slope, and even map the inside of a pyramid from orbit. It’s called Synthetic Aperture Radar (SAR), and a lot of the data
https://x.com/bilawalsidhu/status/2047451145392513105
Wall hacks in real life using augmented reality. New mad science experiment — I used $300 Meta RayBan glasses and an iPhone to recreate EagleEye’s “”see-through-walls”” tracking. The answer was visual positioning tech to build a shared spatial map. Code on GitHub. 0:00 – The
https://x.com/bilawalsidhu/status/2048885939183743385
When your clanker becomes a beast at blender 3d
https://x.com/bilawalsidhu/status/2049636058434859228
FastGHA: Generalized Few-Shot 3D Gaussian Head Avatars with Real-Time Animation”” TL;DR: generates high-quality animatable 3D Gaussian head avatars from few images using a feed-forward transformer and lightweight deformation network
https://x.com/Almorgand/status/2047339475345281341
Pixel3DMM: Versatile Screen-Space Priors for Single-Image 3D Face Reconstruction”” TL;DR: uses vision transformers to predict per-pixel geometry and fit a 3D morphable face model from a single image, achieving strong accuracy across poses and expressions
https://x.com/Almorgand/status/2048785011587858685
Microsoft Presents “”TRELLIS.2″”: An Open-Source, 4B-Parameter, Image-to-3D Model producing up to 1536³ PBR textured assets. Built On Native 3D VAES With 16× Spatial compression, delivering efficient, scalable, high-fidelity asset generation. Ngl, pretty cool!
https://x.com/kimmonismus/status/2049099376476459372
World-R1 | Reinforcing 3D Constraints for Text-to-Video Generation
https://microsoft.github.io/World-R1/





Leave a Reply