Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Recurrent Video Masked Autoencoders

About

We present Recurrent Video Masked-Autoencoders (RVM): a novel video representation learning approach that uses a transformer-based recurrent neural network to aggregate dense image features over time, effectively capturing the spatio-temporal structure of natural video data. RVM learns via an asymmetric masked prediction task requiring only a standard pixel reconstruction objective. This design yields a highly efficient ``generalist'' encoder: RVM achieves competitive performance with state-of-the-art video models (e.g. VideoMAE, V-JEPA) on video-level tasks like action recognition and point/object tracking, while also performing favorably against image models (e.g. DINOv2) on tasks that test geometric and dense spatial understanding. Notably, RVM achieves strong performance in the small-model regime without requiring knowledge distillation, exhibiting up to 30x greater parameter efficiency than competing video masked autoencoders. Moreover, we demonstrate that RVM's recurrent nature allows for stable feature propagation over long temporal horizons with linear computational cost, overcoming some of the limitations of standard spatio-temporal attention-based architectures. Finally, we use qualitative visualizations to highlight that RVM learns rich representations of scene semantics, structure, and motion.

Daniel Zoran, Nikhil Parthasarathy, Yi Yang, Drew A Hudson, Joao Carreira, Andrew Zisserman• 2025

Related benchmarks

TaskDatasetResultRank
Action RecognitionSSV2
Top-1 Acc61.4
93
Action RecognitionKinetics
Top-1 Acc53.1
83
Monocular Depth EstimationScanNet
AbsRel0.97
64
Video Object SegmentationDAVIS
J Mean63.9
58
Human Pose EstimationJHMDB
PCK@0.149.4
12
Video tasksSS v2
Accuracy68.7
11
Video tasksKinetics
Accuracy60
11
Video tasksWaymo
mIoU74.2
11
Spatial TasksDAVIS
J&F Score66
9
Spatial TasksJHMDB
PCK@0.148.4
9
Showing 10 of 14 rows

Other info

Follow for update