MoVie: Visual Model-Based Policy Adaptation for View Generalization
About
Visual Reinforcement Learning (RL) agents trained on limited views face significant challenges in generalizing their learned abilities to unseen views. This inherent difficulty is known as the problem of $\textit{view generalization}$. In this work, we systematically categorize this fundamental problem into four distinct and highly challenging scenarios that closely resemble real-world situations. Subsequently, we propose a straightforward yet effective approach to enable successful adaptation of visual $\textbf{Mo}$del-based policies for $\textbf{Vie}$w generalization ($\textbf{MoVie}$) during test time, without any need for explicit reward signals and any modification during training time. Our method demonstrates substantial advancements across all four scenarios encompassing a total of $\textbf{18}$ tasks sourced from DMControl, xArm, and Adroit, with a relative improvement of $\mathbf{33}$%, $\mathbf{86}$%, and $\mathbf{152}$% respectively. The superior results highlight the immense potential of our approach for real-world robotics applications. Videos are available at https://yangsizhe.github.io/MoVie/ .
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Cup Catch | DMControl Novel view (test) | Reward973.6 | 12 | |
| Finger Spin | DMControl Novel view (test) | Reward917.2 | 12 | |
| Cube Lift | AIRBOT Play CubeLift | Success Rate11.2 | 11 | |
| Continuous Control | DMControl Novel view | Episode Reward770.6 | 8 | |
| Drawer-Open | Maniwhere-inspired Benchmark UR5 | Success Rate5.2 | 8 | |
| Button press | Maniwhere-inspired Benchmark AIRBOT Play | Success Rate10.8 | 8 | |
| Button Press Dex | Maniwhere-inspired Benchmark UR5 | Success Rate9.6 | 8 | |
| Hand Over Dual | Maniwhere-inspired Benchmark Franka | Success Rate11.2 | 8 | |
| Laptop Close | Maniwhere-inspired Benchmark AIRBOT Play | Success Rate4.8 | 8 | |
| Pick & Place Dex | Maniwhere-inspired Benchmark Franka | Success Rate1.6 | 8 |