Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Safe Imitation Learning via Fast Bayesian Reward Inference from Preferences

About

Bayesian reward learning from demonstrations enables rigorous safety and uncertainty analysis when performing imitation learning. However, Bayesian reward learning methods are typically computationally intractable for complex control problems. We propose Bayesian Reward Extrapolation (Bayesian REX), a highly efficient Bayesian reward learning algorithm that scales to high-dimensional imitation learning problems by pre-training a low-dimensional feature encoding via self-supervised tasks and then leveraging preferences over demonstrations to perform fast Bayesian inference. Bayesian REX can learn to play Atari games from demonstrations, without access to the game score and can generate 100,000 samples from the posterior over reward functions in only 5 minutes on a personal laptop. Bayesian REX also results in imitation learning performance that is competitive with or better than state-of-the-art methods that only learn point estimates of the reward function. Finally, Bayesian REX enables efficient high-confidence policy evaluation without having access to samples of the reward function. These high-confidence performance bounds can be used to rank the performance and risk of a variety of evaluation policies and provide a way to detect reward hacking behaviors.

Daniel S. Brown, Russell Coleman, Ravi Srinivasan, Scott Niekum• 2020

Related benchmarks

TaskDatasetResultRank
Offline Reinforcement LearningD4RL hopper-medium-expert
Normalized Score85.1
115
Offline Reinforcement LearningD4RL walker2d-medium-expert
Normalized Score99.62
86
Offline Reinforcement LearningD4RL Medium-Replay Hopper
Normalized Score62
72
Offline Reinforcement LearningD4RL Medium-Replay Walker2d
Normalized Score10.3
34
Offline Reinforcement LearningRobomimic Lift (proficient-human)
Avg Normalized Score96.6
7
Offline Reinforcement LearningRobomimic Can (proficient-human)
Avg Normalized Score63
7
Offline Reinforcement LearningRobomimic Lift multi-human
Avg Normalized Score60.4
7
Offline Reinforcement LearningRobomimic Can multi-human
Avg Normalized Score30.4
7
Showing 8 of 8 rows

Other info

Follow for update