Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Neural Volumetric Memory for Visual Locomotion Control

About

Legged robots have the potential to expand the reach of autonomy beyond paved roads. In this work, we consider the difficult problem of locomotion on challenging terrains using a single forward-facing depth camera. Due to the partial observability of the problem, the robot has to rely on past observations to infer the terrain currently beneath it. To solve this problem, we follow the paradigm in computer vision that explicitly models the 3D geometry of the scene and propose Neural Volumetric Memory (NVM), a geometric memory architecture that explicitly accounts for the SE(3) equivariance of the 3D world. NVM aggregates feature volumes from multiple camera views by first bringing them back to the ego-centric frame of the robot. We test the learned visual-locomotion policy on a physical robot and show that our approach, which explicitly introduces geometric priors during training, offers superior performance than more na\"ive methods. We also include ablation studies and show that the representations stored in the neural volumetric memory capture sufficient geometric information to reconstruct the scene. Our project page with videos is https://rchalyang.github.io/NVM .

Ruihan Yang, Ge Yang, Xiaolong Wang• 2023

Related benchmarks

TaskDatasetResultRank
LocomotionStairs (Simulated)
Traversing Rate85.5
11
LocomotionStages Simulated
Traversing Rate79.9
6
LocomotionStones (Simulated)
Traversing Rate47.4
6
LocomotionObstacles Simulated
Traversing Rate74.1
6
Legged LocomotionReal-world Stages
Distance Moved (m)5.4
2
Legged LocomotionReal-world Stairs
Distance Moved (m)4
2
Legged LocomotionReal-world Obstacles
Distance Moved (m)7.3
2
Showing 7 of 7 rows

Other info

Code

Follow for update