The case for and against fixed step-size: Stochastic approximation algorithms in optimization and machine learning
Theory and application of stochastic approximation (SA) have become increasingly relevant due in part to applications in optimization and reinforcement learning. This paper takes a new look at SA with constant step-size , defined by the recursion, in which and is a Markov chain. The goal is to approximately solve root finding problem , where and has the steady-state distribution of .The following conclusions are obtained under an ergodicity assumption on the Markov chain, compatible assumptions on , and for sufficiently small: The pair process is geometrically ergodic in a topological sense. For every , there is a constant such that for each initial condition. The Polyak-Ruppert-style averaged estimates converge to a limit almost surely and in mean square, which satisfies for an identified non-random . Moreover, the covariance is approximately optimal: The limiting covariance matrix of is approximately minimal in a matricial sense.The two main take-aways for practitioners are application-dependent. It is argued that, in applications to optimization, constant gain algorithms may be preferable even when the objective has multiple local minima; while a vanishing gain algorithm is preferable in applications to reinforcement learning due to the presence of bias.
View on arXiv