ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.00419
41
0

Heavy-Tailed Linear Bandits: Huber Regression with One-Pass Update

1 March 2025
Jing Wang
Yu-Jie Zhang
Peng Zhao
Zhi-Hua Zhou
ArXivPDFHTML
Abstract

We study the stochastic linear bandits with heavy-tailed noise. Two principled strategies for handling heavy-tailed noise, truncation and median-of-means, have been introduced to heavy-tailed bandits. Nonetheless, these methods rely on specific noise assumptions or bandit structures, limiting their applicability to general settings. The recent work [Huang et al.2024] develops a soft truncation method via the adaptive Huber regression to address these limitations. However, their method suffers undesired computational cost: it requires storing all historical data and performing a full pass over these data at each round. In this paper, we propose a \emph{one-pass} algorithm based on the online mirror descent framework. Our method updates using only current data at each round, reducing the per-round computational cost from O~(tlog⁡T)\widetilde{\mathcal{O}}(t \log T)O(tlogT) to O~(1)\widetilde{\mathcal{O}}(1)O(1) with respect to current round ttt and the time horizon TTT, and achieves a near-optimal and variance-aware regret of order O~(dT1−ϵ2(1+ϵ)∑t=1Tνt2+dT1−ϵ2(1+ϵ))\widetilde{\mathcal{O}}\big(d T^{\frac{1-\epsilon}{2(1+\epsilon)}} \sqrt{\sum_{t=1}^T \nu_t^2} + d T^{\frac{1-\epsilon}{2(1+\epsilon)}}\big)O(dT2(1+ϵ)1−ϵ​∑t=1T​νt2​​+dT2(1+ϵ)1−ϵ​) where ddd is the dimension and νt1+ϵ\nu_t^{1+\epsilon}νt1+ϵ​ is the (1+ϵ)(1+\epsilon)(1+ϵ)-th central moment of reward at round ttt.

View on arXiv
@article{wang2025_2503.00419,
  title={ Heavy-Tailed Linear Bandits: Huber Regression with One-Pass Update },
  author={ Jing Wang and Yu-Jie Zhang and Peng Zhao and Zhi-Hua Zhou },
  journal={arXiv preprint arXiv:2503.00419},
  year={ 2025 }
}
Comments on this paper