Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Forecast-MAE: Self-supervised Pre-training for Motion Forecasting with Masked Autoencoders

About

This study explores the application of self-supervised learning (SSL) to the task of motion forecasting, an area that has not yet been extensively investigated despite the widespread success of SSL in computer vision and natural language processing. To address this gap, we introduce Forecast-MAE, an extension of the mask autoencoders framework that is specifically designed for self-supervised learning of the motion forecasting task. Our approach includes a novel masking strategy that leverages the strong interconnections between agents' trajectories and road networks, involving complementary masking of agents' future or history trajectories and random masking of lane segments. Our experiments on the challenging Argoverse 2 motion forecasting benchmark show that Forecast-MAE, which utilizes standard Transformer blocks with minimal inductive bias, achieves competitive performance compared to state-of-the-art methods that rely on supervised learning and sophisticated designs. Moreover, it outperforms the previous self-supervised learning method by a significant margin. Code is available at https://github.com/jchengai/forecast-mae.

Jie Cheng, Xiaodong Mei, Ming Liu• 2023

Related benchmarks

TaskDatasetResultRank
Motion forecastingArgoverse 2 Motion Forecasting Dataset (test)
Miss Rate (K=6)17
90
Trajectory ForecastingNuScenes v1.0 (test)
minADEk1.159
14
Trajectory PredictionnuPlan zero-shot (test)
minADE0.936
9
Trajectory PredictionnuScenes 1.0 (test)
Latency (ms)21.82
9
Multi-agent motion forecastingArgoverse multi-agent 2 (test)
Average minFDE (K=1)3.33
5
Showing 5 of 5 rows

Other info

Follow for update