310

Efficient Exploration for Model-based Reinforcement Learning with Continuous States and Actions

International Conference on Machine Learning (ICML), 2020
Abstract

Balancing exploration and exploitation is crucial in reinforcement learning (RL). In this paper, we study the model-based posterior sampling algorithm in continuous state-action spaces theoretically and empirically. First, we improve the regret bound: with the assumption that reward and transition functions can be modeled as Gaussian Processes with linear kernels, we develop a Bayesian regret bound of O~(H3/2dT)\tilde{O}(H^{3/2}d\sqrt{T}), where HH is the episode length, dd is the dimension of the state-action space, and TT indicates the total time steps. Our bound can be extended to nonlinear cases as well: using linear kernels on the feature representation ϕ\phi, the Bayesian regret bound becomes O~(H3/2dϕT)\tilde{O}(H^{3/2}d_{\phi}\sqrt{T}), where dϕd_\phi is the dimension of the representation space. Moreover, we present MPC-PSRL, a model-based posterior sampling algorithm with model predictive control for action selection. To capture the uncertainty in models and realize posterior sampling, we use Bayesian linear regression on the penultimate layer (the feature representation layer ϕ\phi) of neural networks. Empirical results show that our algorithm achieves the best sample efficiency in benchmark control tasks compared to prior model-based algorithms, and matches the asymptotic performance of model-free algorithms.

View on arXiv
Comments on this paper