Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

CAE: Repurposing the Critic as an Explorer in Deep Reinforcement Learning

About

Exploration remains a fundamental challenge in reinforcement learning, as many existing methods either lack theoretical guarantees or fall short in practical effectiveness. In this paper, we propose CAE, i.e., the Critic as an Explorer, a lightweight approach that repurposes the value networks in standard deep RL algorithms to drive exploration, without introducing additional parameters. CAE leverages multi-armed bandit techniques combined with a tailored scaling strategy, enabling efficient exploration with provable sub-linear regret bounds and strong empirical stability. Remarkably, it is simple to implement, requiring only about 10 lines of code. For complex tasks where learning reliable value networks is difficult, we introduce CAE+, an extension of CAE that incorporates an auxiliary network. CAE+ increases the parameter count by less than 1% while preserving implementation simplicity, adding roughly 10 additional lines of code. Extensive experiments on MuJoCo, MiniHack, and Habitat validate the effectiveness of CAE and CAE+, highlighting their ability to unify theoretical rigor with practical efficiency.

Yexin Li• 2025

Related benchmarks

TaskDatasetResultRank
PointNavHM3D Habitat (test)
Environment Revealed Proportion69
3
ExplorationMiniHack
RPI (%)35.29
1
Showing 2 of 2 rows

Other info

Follow for update