Almost Sure Convergence Rates and Concentration of Stochastic
Approximation and Reinforcement Learning with Markovian Noise
Main:29 Pages
Bibliography:6 Pages
2 Tables
Abstract
This paper establishes the first almost sure convergence rate and the first maximal concentration bound with exponential tails for general contractive stochastic approximation algorithms with Markovian noise. As a corollary, we also obtain convergence rates in . Key to our successes is a novel discretization of the mean ODE of stochastic approximation algorithms using intervals with diminishing (instead of constant) length. As applications, we provide the first almost sure convergence rate for -learning with Markovian samples without count-based learning rates. We also provide the first concentration bound for off-policy temporal difference learning with Markovian samples.
View on arXivComments on this paper
