Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Discriminator-Actor-Critic: Addressing Sample Inefficiency and Reward Bias in Adversarial Imitation Learning

About

We identify two issues with the family of algorithms based on the Adversarial Imitation Learning framework. The first problem is implicit bias present in the reward functions used in these algorithms. While these biases might work well for some environments, they can also lead to sub-optimal behavior in others. Secondly, even though these algorithms can learn from few expert demonstrations, they require a prohibitively large number of interactions with the environment in order to imitate the expert for many real-world applications. In order to address these issues, we propose a new algorithm called Discriminator-Actor-Critic that uses off-policy Reinforcement Learning to reduce policy-environment interaction sample complexity by an average factor of 10. Furthermore, since our reward function is designed to be unbiased, we can apply our algorithm to many problems without making any task-specific adjustments.

Ilya Kostrikov, Kumar Krishna Agrawal, Debidatta Dwibedi, Sergey Levine, Jonathan Tompson• 2018

Related benchmarks

TaskDatasetResultRank
Imitation LearningDeepMind Control Suite image-based
Cartpole Score0.08
6
Single-life task completioncheetah
Average Steps9.92e+4
5
Single-life task completionPointmass
Average Steps1.01e+5
5
Single-life task completionKitchen
Avg Steps1.11e+5
5
Single-life task completionTabletop
Avg Steps8.32e+4
5
Imitation LearningMuJoCo (test)
Hopper Score3.31e+3
5
Imitation LearningDeepMind Control Suite state-based
Cartpole Score13
5
Showing 7 of 7 rows

Other info

Follow for update