ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2310.18701
25
1

Efficient Algorithms for Generalized Linear Bandits with Heavy-tailed Rewards

28 October 2023
Bo Xue
Yimu Wang
Yuanyu Wan
Jinfeng Yi
Lijun Zhang
ArXivPDFHTML
Abstract

This paper investigates the problem of generalized linear bandits with heavy-tailed rewards, whose (1+ϵ)(1+\epsilon)(1+ϵ)-th moment is bounded for some ϵ∈(0,1]\epsilon\in (0,1]ϵ∈(0,1]. Although there exist methods for generalized linear bandits, most of them focus on bounded or sub-Gaussian rewards and are not well-suited for many real-world scenarios, such as financial markets and web-advertising. To address this issue, we propose two novel algorithms based on truncation and mean of medians. These algorithms achieve an almost optimal regret bound of O~(dT11+ϵ)\widetilde{O}(dT^{\frac{1}{1+\epsilon}})O(dT1+ϵ1​), where ddd is the dimension of contextual information and TTT is the time horizon. Our truncation-based algorithm supports online learning, distinguishing it from existing truncation-based approaches. Additionally, our mean-of-medians-based algorithm requires only O(log⁡T)O(\log T)O(logT) rewards and one estimator per epoch, making it more practical. Moreover, our algorithms improve the regret bounds by a logarithmic factor compared to existing algorithms when ϵ=1\epsilon=1ϵ=1. Numerical experimental results confirm the merits of our algorithms.

View on arXiv
Comments on this paper