ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2406.01192
41
1

Sparsity-Agnostic Linear Bandits with Adaptive Adversaries

3 June 2024
Tianyuan Jin
Kyoungseok Jang
Nicolò Cesa-Bianchi
ArXivPDFHTML
Abstract

We study stochastic linear bandits where, in each round, the learner receives a set of actions (i.e., feature vectors), from which it chooses an element and obtains a stochastic reward. The expected reward is a fixed but unknown linear function of the chosen action. We study sparse regret bounds, that depend on the number SSS of non-zero coefficients in the linear reward function. Previous works focused on the case where SSS is known, or the action sets satisfy additional assumptions. In this work, we obtain the first sparse regret bounds that hold when SSS is unknown and the action sets are adversarially generated. Our techniques combine online to confidence set conversions with a novel randomized model selection approach over a hierarchy of nested confidence sets. When SSS is known, our analysis recovers state-of-the-art bounds for adversarial action sets. We also show that a variant of our approach, using Exp3 to dynamically select the confidence sets, can be used to improve the empirical performance of stochastic linear bandits while enjoying a regret bound with optimal dependence on the time horizon.

View on arXiv
Comments on this paper