325

Computing the Bias of Constant-step Stochastic Approximation with Markovian Noise

Neural Information Processing Systems (NeurIPS), 2024
Abstract

We study stochastic approximation algorithms with Markovian noise and constant step-size α\alpha. We develop a method based on infinitesimal generator comparisons to study the bias of the algorithm, which is the expected difference between θn\theta_n -- the value at iteration nn -- and θ\theta^* -- the unique equilibrium of the corresponding ODE. We show that, under some smoothness conditions, this bias is of order O(α)O(\alpha). Furthermore, we show that the time-averaged bias is equal to αV+O(α2)\alpha V + O(\alpha^2), where VV is a constant characterized by a Lyapunov equation, showing that \espθˉnθ+Vα+O(α2)\esp{\bar{\theta}_n} \approx \theta^*+V\alpha + O(\alpha^2), where θˉn=(1/n)k=1nθk\bar{\theta}_n=(1/n)\sum_{k=1}^n\theta_k is the Polyak-Ruppert average. We also show that θˉn\bar{\theta}_n converges with high probability around θ+αV\theta^*+\alpha V. We illustrate how to combine this with Richardson-Romberg extrapolation to derive an iterative scheme with a bias of order O(α2)O(\alpha^2).

View on arXiv
Comments on this paper