Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Conservative Safety Critics for Exploration

About

Safe exploration presents a major challenge in reinforcement learning (RL): when active data collection requires deploying partially trained policies, we must ensure that these policies avoid catastrophically unsafe regions, while still enabling trial and error learning. In this paper, we target the problem of safe exploration in RL by learning a conservative safety estimate of environment states through a critic, and provably upper bound the likelihood of catastrophic failures at every training iteration. We theoretically characterize the tradeoff between safety and policy improvement, show that the safety constraints are likely to be satisfied with high probability during training, derive provable convergence guarantees for our approach, which is no worse asymptotically than standard RL, and demonstrate the efficacy of the proposed approach on a suite of challenging navigation, manipulation, and locomotion tasks. Empirically, we show that the proposed approach can achieve competitive task performance while incurring significantly lower catastrophic failure rates during training than prior methods. Videos are at this url https://sites.google.com/view/conservative-safety-critics/home

Homanga Bharadhwaj, Aviral Kumar, Nicholas Rhinehart, Sergey Levine, Florian Shkurti, Animesh Garg• 2020

Related benchmarks

TaskDatasetResultRank
FetchReachGymnasium Robotics
Reward4.78
16
Goal1Safety Gymnasium
Reward7.21
16
Goal2Safety Gymnasium
Reward9.44
16
HalfCheetahMujoco
Reward8.66
16
Button1Safety Gymnasium
Reward5.48
16
Button2Safety Gymnasium
Reward6.45
16
Button navigationSafety Gymnasium Button1 v0 (test)
Success Rate70
8
Button navigationSafety Gymnasium Button2 v0 (test)
Success Rate80
8
Robotic reachingGymnasium Robotics FetchReach v0 (test)
Success Rate100
8
Goal AchievementSafety Gymnasium Goal1 v0 (test)
Success Rate100
8
Showing 10 of 11 rows

Other info

Follow for update