Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Information Directed Reward Learning for Reinforcement Learning

About

For many reinforcement learning (RL) applications, specifying a reward is difficult. This paper considers an RL setting where the agent obtains information about the reward only by querying an expert that can, for example, evaluate individual states or provide binary preferences over trajectories. From such expensive feedback, we aim to learn a model of the reward that allows standard RL algorithms to achieve high expected returns with as few expert queries as possible. To this end, we propose Information Directed Reward Learning (IDRL), which uses a Bayesian model of the reward and selects queries that maximize the information gain about the difference in return between plausibly optimal policies. In contrast to prior active reward learning methods designed for specific types of queries, IDRL naturally accommodates different query types. Moreover, it achieves similar or better performance with significantly fewer queries by shifting the focus from reducing the reward approximation error to improving the policy induced by the reward model. We support our findings with extensive evaluations in multiple environments and with different query types.

David Lindner, Matteo Turchetta, Sebastian Tschiatschek, Kamil Ciosek, Andreas Krause• 2021

Related benchmarks

TaskDatasetResultRank
Reinforcement LearningAtari Breakout
Mean Return153.7
23
Reinforcement LearningAtari 2600 Qbert
Score8.29e+3
20
Reinforcement LearningAtari Pong
Mean Episode Return15.3
19
Reinforcement LearningAtari 2600 Seaquest
Average Score2.94e+3
12
Offline Reinforcement LearningAntMaze
Success Rate (umaze)85.5
5
Reinforcement LearningAtari Asterix
Score357.4
5
Offline Preference-Based Reinforcement LearningMeta-World
Lever Pull Success Rate33.1
5
Showing 7 of 7 rows

Other info

Follow for update