408
v1v2v3 (latest)

ETGL-DDPG: A Deep Deterministic Policy Gradient Algorithm for Sparse Reward Continuous Control

Main:13 Pages
17 Figures
Bibliography:6 Pages
5 Tables
Appendix:12 Pages
Abstract

We consider deep deterministic policy gradient (DDPG) in the context of reinforcement learning with sparse rewards. To enhance exploration, we introduce a search procedure, \emph{ϵt{\epsilon}{t}-greedy}, which generates exploratory options for exploring less-visited states. We prove that search using ϵt\epsilon t-greedy has polynomial sample complexity under mild MDP assumptions. To more efficiently use the information provided by rewarded transitions, we develop a new dual experience replay buffer framework, \emph{GDRB}, and implement \emph{longest n-step returns}. The resulting algorithm, \emph{ETGL-DDPG}, integrates all three techniques: \bm{ϵt\epsilon t}-greedy, \textbf{G}DRB, and \textbf{L}ongest nn-step, into DDPG. We evaluate ETGL-DDPG on standard benchmarks and demonstrate that it outperforms DDPG, as well as other state-of-the-art methods, across all tested sparse-reward continuous environments. Ablation studies further highlight how each strategy individually enhances the performance of DDPG in this setting.

View on arXiv
Comments on this paper