ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2106.15662
47
2

Exponential Weights Algorithms for Selective Learning

29 June 2021
Mingda Qiao
Gregory Valiant
ArXiv (abs)PDFHTML
Abstract

We study the selective learning problem introduced by Qiao and Valiant (2019), in which the learner observes nnn labeled data points one at a time. At a time of its choosing, the learner selects a window length www and a model ℓ^\hat\ellℓ^ from the model class L\mathcal{L}L, and then labels the next www data points using ℓ^\hat\ellℓ^. The excess risk incurred by the learner is defined as the difference between the average loss of ℓ^\hat\ellℓ^ over those www data points and the smallest possible average loss among all models in L\mathcal{L}L over those www data points. We give an improved algorithm, termed the hybrid exponential weights algorithm, that achieves an expected excess risk of O((log⁡log⁡∣L∣+log⁡log⁡n)/log⁡n)O((\log\log|\mathcal{L}| + \log\log n)/\log n)O((loglog∣L∣+loglogn)/logn). This result gives a doubly exponential improvement in the dependence on ∣L∣|\mathcal{L}|∣L∣ over the best known bound of O(∣L∣/log⁡n)O(\sqrt{|\mathcal{L}|/\log n})O(∣L∣/logn​). We complement the positive result with an almost matching lower bound, which suggests the worst-case optimality of the algorithm. We also study a more restrictive family of learning algorithms that are bounded-recall in the sense that when a prediction window of length www is chosen, the learner's decision only depends on the most recent www data points. We analyze an exponential weights variant of the ERM algorithm in Qiao and Valiant (2019). This new algorithm achieves an expected excess risk of O(log⁡∣L∣/log⁡n)O(\sqrt{\log |\mathcal{L}|/\log n})O(log∣L∣/logn​), which is shown to be nearly optimal among all bounded-recall learners. Our analysis builds on a generalized version of the selective mean prediction problem in Drucker (2013); Qiao and Valiant (2019), which may be of independent interest.

View on arXiv
Comments on this paper