Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Video World Models with Long-term Spatial Memory

About

Emerging world models autoregressively generate video frames in response to actions, such as camera movements and text prompts, among other control signals. Due to limited temporal context window sizes, these models often struggle to maintain scene consistency during revisits, leading to severe forgetting of previously generated environments. Inspired by the mechanisms of human memory, we introduce a novel framework to enhancing long-term consistency of video world models through a geometry-grounded long-term spatial memory. Our framework includes mechanisms to store and retrieve information from the long-term spatial memory and we curate custom datasets to train and evaluate world models with explicitly stored 3D memory mechanisms. Our evaluations show improved quality, consistency, and context length compared to relevant baselines, paving the way towards long-term consistent world generation.

Tong Wu, Shuai Yang, Ryan Po, Yinghao Xu, Ziwei Liu, Dahua Lin, Gordon Wetzstein• 2025

Related benchmarks

TaskDatasetResultRank
Memory Retrieval ConsistencyMosaicMem Memory Retrieval (test)
SSIM65
12
Camera Motion ControlMosaicMem Camera Control (Dedicated Evaluation Set)
Rotational Error1.5
12
Video GenerationMosaicMem Dedicated Evaluation Set Overall Generation
FID75.83
12
Motion Dynamics ModelingMosaicMem Motion Dynamics (Dedicated Evaluation Set)
Dynamic Score1.41
12
Video GenerationRealEstate10K and DL3DV partial-revisit (evaluation)
Total Quality Score76.85
11
I2V Camera ControlDL3DV (test)
RRE1.21
10
Long Video GenerationDL3DV-Evaluation (test)
SSIM0.383
8
Long Video GenerationTanks&Temples (test)
SSIM38.3
8
3D Scene GenerationDL3DV (test)
LPIPS (P)0.419
7
3D Scene GenerationTanks&Temples (test)
LPIPS (Perceptual)0.412
7
Showing 10 of 10 rows

Other info

Follow for update