Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Stabilizing Transformers for Reinforcement Learning

About

Owing to their ability to both effectively integrate information over long time horizons and scale to massive amounts of data, self-attention architectures have recently shown breakthrough success in natural language processing (NLP), achieving state-of-the-art results in domains such as language modeling and machine translation. Harnessing the transformer's ability to process long time horizons of information could provide a similar performance boost in partially observable reinforcement learning (RL) domains, but the large-scale transformers used in NLP have yet to be successfully applied to the RL setting. In this work we demonstrate that the standard transformer architecture is difficult to optimize, which was previously observed in the supervised learning setting but becomes especially pronounced with RL objectives. We propose architectural modifications that substantially improve the stability and learning speed of the original Transformer and XL variant. The proposed architecture, the Gated Transformer-XL (GTrXL), surpasses LSTMs on challenging memory environments and achieves state-of-the-art results on the multi-task DMLab-30 benchmark suite, exceeding the performance of an external memory architecture. We show that the GTrXL, trained using the same losses, has stability and performance that consistently matches or exceeds a competitive LSTM baseline, including on more reactive tasks where memory is less critical. GTrXL offers an easy-to-train, simple-to-implement but substantially more expressive architectural alternative to the standard multi-layer LSTM ubiquitously used for RL agents in partially observable environments.

Emilio Parisotto, H. Francis Song, Jack W. Rae, Razvan Pascanu, Caglar Gulcehre, Siddhant M. Jayakumar, Max Jaderberg, Raphael Lopez Kaufman, Aidan Clark, Seb Noury, Matthew M. Botvinick, Nicolas Heess, Raia Hadsell• 2019

Related benchmarks

TaskDatasetResultRank
Command tracking recovery after pushFlat terrain task
Recovery Steps6.7
28
Humanoid LocomotionHumanoid Randomized Task (OOD Sweep)
Reward-4.65
24
Humanoid LocomotionFlat In-distribution (deterministic evaluation)
Cumulative Reward5.02
4
Humanoid LocomotionPush In-distribution (deterministic evaluation)
Cumulative Reward4.83
4
Humanoid LocomotionRandomized In-distribution (deterministic evaluation)
Cumulative Reward4.77
4
Randomized control taskCombined-shift L0/ID
Reward4.67
4
Randomized control taskCombined-shift L1
Reward4.73
4
Randomized control taskCombined-shift L2
Reward4.72
4
Robot Control RobustnessRandomized Task Combined-shift sweep
ID Reward4.77
4
Robot LocomotionRobot Locomotion In-Distribution
Mean Reward4.77
4
Showing 10 of 18 rows

Other info

Follow for update