ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2009.03017
81
19
v1v2v3v4v5 (latest)

Non-exponentially weighted aggregation: regret bounds for unbounded loss functions

7 September 2020
Pierre Alquier
ArXiv (abs)PDFHTML
Abstract

We tackle the problem of online optimization with a general, possibly unbounded, loss function. It is well known that the exponentially weighted aggregation strategy (EWA) leads to a regret in T\sqrt{T}T​ after TTT steps, under the assumption that the loss is bounded. The online gradient algorithm (OGA) has a regret in T\sqrt{T}T​ when the loss is convex and Lipschitz. In this paper, we study a generalized aggregation strategy, where the weights do no longer necessarily depend exponentially on the losses. Our strategy can be interpreted as the minimization of the expected losses plus a penalty term. When the penalty term is the Kullback-Leibler divergence, we obtain EWA as a special case, but using alternative divergences lead to a regret bounds for unbounded, not necessarily convex losses. However, the cost is a worst regret bound in some cases.

View on arXiv
Comments on this paper