SEA-Nav: Efficient Policy Learning for Safe and Agile Quadruped Navigation in Cluttered Environments
About
Efficiently training quadruped robot navigation in densely cluttered environments remains a significant challenge. Existing methods are either limited by a lack of safety and agility in simple obstacle distributions or suffer from slow locomotion in complex environments, often requiring excessively long training phases. To this end, we propose SEA-Nav (Safe, Efficient, and Agile Navigation), a reinforcement learning framework for quadruped navigation. Within diverse and dense obstacle environments, a differentiable control barrier function (CBF)-based shield constraints the navigation policy to output safe velocity commands. An adaptive collision replay mechanism and hazardous exploration rewards are introduced to increase the probability of learning from critical experiences, guiding efficient exploration and exploitation. Finally, kinematic action constraints are incorporated to ensure safe velocity commands, facilitating successful physical deployment. To the best of our knowledge, this is the first approach that achieves highly challenging quadruped navigation in the real world with minute-level training time.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Obstacle Avoidance Navigation | Isaac Gym Navigation Easy | Success Rate100 | 7 | |
| Obstacle Avoidance Navigation | Isaac Gym Navigation Medium | Success Rate97 | 7 | |
| Obstacle Avoidance Navigation | Isaac Gym Navigation Hard | Success Rate (SR)90 | 7 | |
| Robot navigation | Cluttered Room | Success Rate (SR)100 | 6 | |
| Robot navigation | Dynamic Obstacle | Success Rate100 | 6 | |
| Robot navigation | Obstacle Course | Success Rate100 | 6 | |
| Robot navigation | S-Blend Track | Success Rate (SR)100 | 6 |