Dual Averaging on Compactly-Supported Distributions And Application to
No-Regret Learning on a Continuum
Abstract
We consider an online learning problem on a continuum. A decision maker is given a compact feasible set , and is faced with the following sequential problem: at iteration~, the decision maker chooses a distribution , then a loss function is revealed, and the decision maker incurs expected loss . We view the problem as an online convex optimization problem on the space of Lebesgue-continnuous distributions on . We prove a general regret bound for the Dual Averaging method on , then prove that dual averaging with -potentials (a class of strongly convex regularizers) achieves sublinear regret when is uniformly fat (a condition weaker than convexity).
View on arXivComments on this paper
