A General Control-Theoretic Approach for Reinforcement Learning: Theory and Algorithms
About
We devise a control-theoretic reinforcement learning approach to support direct learning of the optimal policy. We establish various theoretical properties of our approach, such as convergence and optimality of our analog of the Bellman operator and Q-learning, a new control-policy-variable gradient theorem, and a specific gradient ascent algorithm based on this theorem within the context of a specific control-theoretic framework. We empirically evaluate the performance of our control theoretic approach on several classical reinforcement learning tasks, demonstrating significant improvements in solution quality, sample complexity, and running time of our approach over state-of-the-art methods.
Weiqin Chen, Mark S. Squillante, Chai Wah Wu, Santiago Paternain• 2024
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Reinforcement Learning | LunarLanderContinuous v2 | Mean Reward280.4 | 59 | |
| Reinforcement Learning | CartPole v0 | Mean Score200 | 48 | |
| Reinforcement Learning | MountainCarContinuous v0 | Average Agent Reward93.63 | 48 | |
| Reinforcement Learning Control | Pendulum v1 | Mean Score1.38e+3 | 40 |
Showing 4 of 4 rows