Skip to content
Back to articles

DeepMind's D4RT Rebuilds Moving 3D Scenes 300x Faster

March 7, 2026·4 min read·860 words
AIComputer VisionVideo Summary
4D scene reconstruction showing point cloud tracking of dynamic objects
Image: Screenshot from YouTube.

Key insights

  • D4RT replaces a patchwork of specialized AI models with one transformer that handles depth, motion, and camera angles at once
  • The system predicts where objects are even when hidden behind other objects, solving a long-standing computer vision problem
  • Speed gains come from parallel processing where each point is reconstructed independently, with no inter-communication needed
SourceYouTube
Published March 7, 2026
Two Minute Papers
Two Minute Papers
Hosts:Károly Zsolnai-Fehér

This article is a summary of How DeepMind's New AI Predicts What It Cannot See. Watch the video

Read this article in norsk


In Brief

A new research paper from Google DeepMind, University College London, and the University of Oxford introduces D4RT, a system that reconstructs moving 3D scenes from ordinary video. Dr. Károly Zsolnai-Fehér of Two Minute Papers describes it as a major step forward: one AI model replaces a patchwork of specialized systems, runs up to 300 times faster, and can even predict where objects are when they disappear behind other things.

300x
faster than previous methods
1
transformer replaces multiple models
4D
3D + time reconstruction

The central claim

Zsolnai-Fehér argues that D4RT represents a fundamental shift in how AI understands video. The name stands for 4D reconstruction with transformers, where the four dimensions are the three spatial dimensions plus time (0:48).

The core idea: feed in a video, get out a 3D point cloud (a collection of dots representing surfaces) that moves and changes over time (1:05). Unlike a static 3D scan, this captures dynamic scenes like judo matches or people walking through rooms.

What makes it different

Previous approaches required multiple specialized AI models, one for depth (how far away things are), one for motion, and one for camera angles. These had to be glued together and then negotiate agreement through a slow process called test-time optimization (2:20).

D4RT replaces all of this with a single transformer (the same type of AI architecture behind ChatGPT and Claude). One model handles depth, motion, and camera position at the same time (2:52).

Seeing the invisible

The most striking claim: D4RT can track objects even when they are hidden behind other objects, a problem called occlusion (3:18). If a chair leg disappears behind a sofa, the system remembers where it was before and predicts where it will reappear. It reconstructs what it cannot see by reasoning about the full video sequence rather than individual frames (8:10).


Why it is so fast

The speed advantage comes from the architecture. Zsolnai-Fehér uses an analogy: imagine a master carpenter who understands the whole scene (the encoder) directing individual elves (the decoder) to each place one screw (5:57).

The key insight: the elves don't need to talk to each other. Each point in the scene is reconstructed independently, which means the work can be split across as many processors as you have (6:59). The technique is fully parallelizable, which is the main reason it achieves up to 300x speed improvements (4:02).

An additional trick recovers fine detail: the original high-resolution video pixels are fed back into the decoder, letting it reconstruct details finer than the AI's own internal representation (7:27).


How to interpret these claims

Zsolnai-Fehér is upfront about limitations, which strengthens the presentation. Three weaknesses are worth noting.

Point clouds are "unintelligent" data. The output is a collection of dots. You cannot 3D print it or use it for physics simulations without an extra step to convert it into a mesh (a surface made of connected triangles) (5:03).

It is not photorealistic. Traditional 3D meshes and Gaussian splats (a newer technique using blurry blobs to represent scenes) still produce more visually realistic results. D4RT focuses on geometric accuracy, not pretty rendering (5:23).

Editing is difficult. Without the structured faces of a mesh, you cannot open the result in a tool like Blender and sculpt it like digital clay (5:36).

The "300x faster" claim also deserves context. The comparison is against test-time optimization methods, which are notoriously slow. Against real-time rendering pipelines used in games, the comparison would look different. The paper itself likely provides more precise benchmarks.


Practical implications

For game and film studios

D4RT could speed up the process of turning real-world footage into 3D assets. The point cloud output would need conversion to meshes for production use, but the geometric and motion data could serve as a starting point.

For robotics and autonomous systems

Understanding moving 3D scenes from video is critical for robots and self-driving vehicles. The ability to predict where objects are, even when hidden, addresses a real safety concern.


Glossary

TermDefinition
4D reconstructionRebuilding a 3D scene that changes over time from video footage. The four dimensions are width, height, depth, and time.
Point cloudA collection of 3D points representing the surfaces of objects. Like a connect-the-dots puzzle before the dots are connected.
TransformerA type of AI architecture that processes data in parallel. The same technology behind language models like ChatGPT and Claude.
OcclusionWhen an object is hidden behind another object. A common challenge in computer vision.
Test-time optimizationA slow process where multiple AI models negotiate with each other at runtime to produce consistent results.
Gaussian splatA 3D representation using overlapping blurry blobs that can render photorealistic scenes. A newer alternative to traditional meshes.
3D meshA surface made of connected triangles, the standard format for 3D objects in games and movies.
Encoder-decoderA two-part AI architecture where the encoder understands the input and the decoder produces the output.
ParallelizableCan be split into independent tasks that run at the same time, making it faster with more processors.

Sources and resources