UVIP: Model-Free Approach to Evaluate Reinforcement Learning Algorithms
- OffRL
Main:14 Pages
4 Figures
Bibliography:3 Pages
Appendix:9 Pages
Abstract
Policy evaluation is an important instrument for the comparison of different algorithms in Reinforcement Learning (RL). However, even a precise knowledge of the value function corresponding to a policy does not provide reliable information on how far the policy is from the optimal one. We present a novel model-free upper value iteration procedure ({\sf UVIP}) that allows us to estimate the suboptimality gap from above and to construct confidence intervals for \(V^\star\). Our approach relies on upper bounds to the solution of the Bellman optimality equation via the martingale approach. We provide theoretical guarantees for {\sf UVIP} under general assumptions and illustrate its performance on a number of benchmark RL problems.
View on arXivComments on this paper
