ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2202.06151
  4. Cited By
Corralling a Larger Band of Bandits: A Case Study on Switching Regret
  for Linear Bandits

Corralling a Larger Band of Bandits: A Case Study on Switching Regret for Linear Bandits

12 February 2022
Haipeng Luo
Mengxiao Zhang
Peng Zhao
Zhi-Hua Zhou
ArXivPDFHTML

Papers citing "Corralling a Larger Band of Bandits: A Case Study on Switching Regret for Linear Bandits"

15 / 15 papers shown
Title
Offline-to-online hyperparameter transfer for stochastic bandits
Dravyansh Sharma
Arun Sai Suggala
OffRL
25
1
0
06 Jan 2025
A Model Selection Approach for Corruption Robust Reinforcement Learning
A Model Selection Approach for Corruption Robust Reinforcement Learning
Chen-Yu Wei
Christoph Dann
Julian Zimmert
80
44
0
31 Dec 2024
An Equivalence Between Static and Dynamic Regret Minimization
An Equivalence Between Static and Dynamic Regret Minimization
Andrew Jacobsen
Francesco Orabona
29
1
0
03 Jun 2024
Universal Online Convex Optimization with $1$ Projection per Round
Universal Online Convex Optimization with 111 Projection per Round
Wenhao Yang
Yibo Wang
Peng Zhao
Lijun Zhang
34
2
0
30 May 2024
Online Linear Regression in Dynamic Environments via Discounting
Online Linear Regression in Dynamic Environments via Discounting
Andrew Jacobsen
Ashok Cutkosky
36
5
0
29 May 2024
Bypassing the Simulator: Near-Optimal Adversarial Linear Contextual
  Bandits
Bypassing the Simulator: Near-Optimal Adversarial Linear Contextual Bandits
Haolin Liu
Chen-Yu Wei
Julian Zimmert
30
9
0
02 Sep 2023
Anytime Model Selection in Linear Bandits
Anytime Model Selection in Linear Bandits
Parnian Kassraie
N. Emmenegger
Andreas Krause
Aldo Pacchiano
39
2
0
24 Jul 2023
Unconstrained Online Learning with Unbounded Losses
Unconstrained Online Learning with Unbounded Losses
Andrew Jacobsen
Ashok Cutkosky
32
13
0
08 Jun 2023
A Blackbox Approach to Best of Both Worlds in Bandits and Beyond
A Blackbox Approach to Best of Both Worlds in Bandits and Beyond
Christoph Dann
Chen-Yu Wei
Julian Zimmert
17
22
0
20 Feb 2023
Optimal Stochastic Non-smooth Non-convex Optimization through
  Online-to-Non-convex Conversion
Optimal Stochastic Non-smooth Non-convex Optimization through Online-to-Non-convex Conversion
Ashok Cutkosky
Harsh Mehta
Francesco Orabona
33
32
0
07 Feb 2023
Adapting to Continuous Covariate Shift via Online Density Ratio
  Estimation
Adapting to Continuous Covariate Shift via Online Density Ratio Estimation
Yu-Jie Zhang
Zhenyu Zhang
Peng Zhao
Masashi Sugiyama
OOD
14
11
0
06 Feb 2023
Unconstrained Dynamic Regret via Sparse Coding
Unconstrained Dynamic Regret via Sparse Coding
Zhiyu Zhang
Ashok Cutkosky
I. Paschalidis
21
7
0
31 Jan 2023
Dynamic Regret of Online Markov Decision Processes
Dynamic Regret of Online Markov Decision Processes
Peng Zhao
Longfei Li
Zhi-Hua Zhou
OffRL
22
17
0
26 Aug 2022
Adversarial Bandits against Arbitrary Strategies
Adversarial Bandits against Arbitrary Strategies
Jung-hun Kim
Se-Young Yun
39
0
0
30 May 2022
Non-stationary Online Learning with Memory and Non-stochastic Control
Non-stationary Online Learning with Memory and Non-stochastic Control
Peng Zhao
Yu-Hu Yan
Yu-Xiang Wang
Zhi-Hua Zhou
33
47
0
07 Feb 2021
1