ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2011.11222
79
20
v1v2 (latest)

Improved Confidence Bounds for the Linear Logistic Model and Applications to Linear Bandits

23 November 2020
Kwang-Sung Jun
Lalit P. Jain
Blake Mason
ArXiv (abs)PDFHTML
Abstract

We propose improved fixed-design confidence bounds for the linear logistic model. Our bounds significantly improve upon the state-of-the-art bounds of Li et al. (2017) by leveraging the self-concordance of the logistic loss inspired by Faury et al. (2020). Specifically, our confidence width does not scale with the problem dependent parameter 1/κ1/\kappa1/κ, where κ\kappaκ is the worst-case variance of an arm reward. At worse, κ\kappaκ scales exponentially with the norm of the unknown linear parameter θ∗\theta^*θ∗. Instead, our bound scales directly on the local variance induced by θ∗\theta^*θ∗. We present two applications of our novel bounds on two logistic bandit problems: regret minimization and pure exploration. Our analysis shows that the new confidence bounds improve upon previous state-of-the-art performance guarantees.

View on arXiv
Comments on this paper