84

Dual Averaging on Compactly-Supported Distributions And Application to No-Regret Learning on a Continuum

Abstract

We consider an online learning problem on a continuum. A decision maker is given a compact feasible set SS, and is faced with the following sequential problem: at iteration~tt, the decision maker chooses a distribution x(t)Δ(S)x^{(t)} \in \Delta(S), then a loss function (t):SR+\ell^{(t)} : S \to \mathbb{R}_+ is revealed, and the decision maker incurs expected loss (t),x(t)=Esx(t)(t)(s)\langle \ell^{(t)}, x^{(t)} \rangle = \mathbb{E}_{s \sim x^{(t)}} \ell^{(t)}(s). We view the problem as an online convex optimization problem on the space Δ(S)\Delta(S) of Lebesgue-continnuous distributions on SS. We prove a general regret bound for the Dual Averaging method on L2(S)L^2(S), then prove that dual averaging with ω\omega-potentials (a class of strongly convex regularizers) achieves sublinear regret when SS is uniformly fat (a condition weaker than convexity).

View on arXiv
Comments on this paper