82
0

Small steps no more: Global convergence of stochastic gradient bandits for arbitrary learning rates

Abstract

We provide a new understanding of the stochastic gradient bandit algorithm by showing that it converges to a globally optimal policy almost surely using \emph{any} constant learning rate. This result demonstrates that the stochastic gradient algorithm continues to balance exploration and exploitation appropriately even in scenarios where standard smoothness and noise control assumptions break down. The proofs are based on novel findings about action sampling rates and the relationship between cumulative progress and noise, and extend the current understanding of how simple stochastic gradient methods behave in bandit settings.

View on arXiv
@article{mei2025_2502.07141,
  title={ Small steps no more: Global convergence of stochastic gradient bandits for arbitrary learning rates },
  author={ Jincheng Mei and Bo Dai and Alekh Agarwal and Sharan Vaswani and Anant Raj and Csaba Szepesvari and Dale Schuurmans },
  journal={arXiv preprint arXiv:2502.07141},
  year={ 2025 }
}
Comments on this paper