Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Multi-View Masked World Models for Visual Robotic Manipulation

About

Visual robotic manipulation research and applications often use multiple cameras, or views, to better perceive the world. How else can we utilize the richness of multi-view data? In this paper, we investigate how to learn good representations with multi-view data and utilize them for visual robotic manipulation. Specifically, we train a multi-view masked autoencoder which reconstructs pixels of randomly masked viewpoints and then learn a world model operating on the representations from the autoencoder. We demonstrate the effectiveness of our method in a range of scenarios, including multi-view control and single-view control with auxiliary cameras for representation learning. We also show that the multi-view masked autoencoder trained with multiple randomized viewpoints enables training a policy with strong viewpoint randomization and transferring the policy to solve real-robot tasks without camera calibration and an adaptation procedure. Video demonstrations are available at: https://sites.google.com/view/mv-mwm.

Younggyo Seo, Junsu Kim, Stephen James, Kimin Lee, Jinwoo Shin, Pieter Abbeel• 2023

Related benchmarks

TaskDatasetResultRank
Cube LiftAIRBOT Play CubeLift
Success Rate59.6
11
Button Press DexManiwhere-inspired Benchmark UR5
Success Rate75.2
8
Laptop CloseManiwhere-inspired Benchmark AIRBOT Play
Success Rate66.8
8
ReachManiwhere-inspired Benchmark AIRBOT Play
Success Rate76.8
8
Reach DexManiwhere-inspired Benchmark UR5
Success Rate76.4
8
Button pressManiwhere-inspired Benchmark AIRBOT Play
Success Rate61.2
8
Drawer-OpenManiwhere-inspired Benchmark UR5
Success Rate47.2
8
Hand Over DualManiwhere-inspired Benchmark Franka
Success Rate35.2
8
Pick & Place DexManiwhere-inspired Benchmark Franka
Success Rate29.6
8
Pick-&-PlaceManiwhere-inspired Benchmark AIRBOT Play
Success Rate30.8
8
Showing 10 of 10 rows

Other info

Follow for update