ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2302.09739
  4. Cited By
A Blackbox Approach to Best of Both Worlds in Bandits and Beyond

A Blackbox Approach to Best of Both Worlds in Bandits and Beyond

Annual Conference Computational Learning Theory (COLT), 2023
20 February 2023
Christoph Dann
Chen-Yu Wei
Julian Zimmert
ArXiv (abs)PDFHTML

Papers citing "A Blackbox Approach to Best of Both Worlds in Bandits and Beyond"

26 / 26 papers shown
Adapting to Stochastic and Adversarial Losses in Episodic MDPs with Aggregate Bandit Feedback
Adapting to Stochastic and Adversarial Losses in Episodic MDPs with Aggregate Bandit Feedback
Shinji Ito
Kevin Jamieson
Haipeng Luo
Arnab Maiti
Taira Tsuchiya
OffRL
253
2
0
20 Oct 2025
FLEET: Formal Language-Grounded Scheduling for Heterogeneous Robot Teams
FLEET: Formal Language-Grounded Scheduling for Heterogeneous Robot Teams
Corban G. Rivera
Grayson Byrd
Meghan Booker
Bethany Kemp
Allison Gaines
Emma Holmes
James Uplinger
Celso M. De Melo
D. Handelman
171
0
0
08 Oct 2025
Efficient Best-of-Both-Worlds Algorithms for Contextual Combinatorial Semi-Bandits
Efficient Best-of-Both-Worlds Algorithms for Contextual Combinatorial Semi-Bandits
Mengmeng Li
Philipp Schneider
Jelisaveta Aleksić
Daniel Kuhn
123
1
0
26 Aug 2025
Revisiting Follow-the-Perturbed-Leader with Unbounded Perturbations in Bandit Problems
Revisiting Follow-the-Perturbed-Leader with Unbounded Perturbations in Bandit Problems
Jongyeong Lee
Junya Honda
Shinji Ito
Min-hwan Oh
226
2
0
26 Aug 2025
A Near-optimal, Scalable and Parallelizable Framework for Stochastic Bandits Robust to Adversarial Corruptions and Beyond
A Near-optimal, Scalable and Parallelizable Framework for Stochastic Bandits Robust to Adversarial Corruptions and Beyond
Zicheng Hu
Cheng Chen
449
0
0
11 Feb 2025
Tracking Most Significant Shifts in Infinite-Armed Bandits
Tracking Most Significant Shifts in Infinite-Armed Bandits
Joe Suk
Jung-hun Kim
383
1
0
31 Jan 2025
A Model Selection Approach for Corruption Robust Reinforcement Learning
A Model Selection Approach for Corruption Robust Reinforcement LearningInternational Conference on Algorithmic Learning Theory (ALT), 2021
Chen-Yu Wei
Christoph Dann
Julian Zimmert
393
51
0
31 Dec 2024
How Does Variance Shape the Regret in Contextual Bandits?
How Does Variance Shape the Regret in Contextual Bandits?Neural Information Processing Systems (NeurIPS), 2024
Zeyu Jia
Jian Qian
Alexander Rakhlin
Chen-Yu Wei
500
10
0
16 Oct 2024
Corruption-Robust Linear Bandits: Minimax Optimality and Gap-Dependent
  Misspecification
Corruption-Robust Linear Bandits: Minimax Optimality and Gap-Dependent MisspecificationNeural Information Processing Systems (NeurIPS), 2024
Haolin Liu
Artin Tajdini
Andrew Wagenmaker
Chen-Yu Wei
510
3
0
10 Oct 2024
uniINF: Best-of-Both-Worlds Algorithm for Parameter-Free Heavy-Tailed MABs
uniINF: Best-of-Both-Worlds Algorithm for Parameter-Free Heavy-Tailed MABsInternational Conference on Learning Representations (ICLR), 2024
Yu Chen
Jiatai Huang
Yan Dai
Longbo Huang
458
6
0
04 Oct 2024
A Simple and Adaptive Learning Rate for FTRL in Online Learning with
  Minimax Regret of $Θ(T^{2/3})$ and its Application to
  Best-of-Both-Worlds
A Simple and Adaptive Learning Rate for FTRL in Online Learning with Minimax Regret of Θ(T2/3)Θ(T^{2/3})Θ(T2/3) and its Application to Best-of-Both-Worlds
Taira Tsuchiya
Shinji Ito
481
4
0
30 May 2024
LC-Tsallis-INF: Generalized Best-of-Both-Worlds Linear Contextual Bandits
LC-Tsallis-INF: Generalized Best-of-Both-Worlds Linear Contextual Bandits
Masahiro Kato
Shinji Ito
587
2
0
05 Mar 2024
Adaptive Learning Rate for Follow-the-Regularized-Leader: Competitive
  Analysis and Best-of-Both-Worlds
Adaptive Learning Rate for Follow-the-Regularized-Leader: Competitive Analysis and Best-of-Both-Worlds
Shinji Ito
Taira Tsuchiya
Junya Honda
488
9
0
01 Mar 2024
Information Capacity Regret Bounds for Bandits with Mediator Feedback
Information Capacity Regret Bounds for Bandits with Mediator Feedback
Khaled Eldowa
Nicolò Cesa-Bianchi
Alberto Maria Metelli
Marcello Restelli
254
3
0
15 Feb 2024
Exploration by Optimization with Hybrid Regularizers: Logarithmic Regret
  with Adversarial Robustness in Partial Monitoring
Exploration by Optimization with Hybrid Regularizers: Logarithmic Regret with Adversarial Robustness in Partial Monitoring
Taira Tsuchiya
Shinji Ito
Junya Honda
308
3
0
13 Feb 2024
Efficient Contextual Bandits with Uninformed Feedback Graphs
Efficient Contextual Bandits with Uninformed Feedback Graphs
Mengxiao Zhang
Yuheng Zhang
Haipeng Luo
Paul Mineiro
245
5
0
12 Feb 2024
Best-of-Both-Worlds Linear Contextual Bandits
Best-of-Both-Worlds Linear Contextual Bandits
Masahiro Kato
Shinji Ito
309
2
0
27 Dec 2023
Best-of-Both-Worlds Algorithms for Linear Contextual Bandits
Best-of-Both-Worlds Algorithms for Linear Contextual Bandits
Yuko Kuroki
Alberto Rumi
Taira Tsuchiya
Fabio Vitale
Nicolò Cesa-Bianchi
338
13
0
24 Dec 2023
Towards Optimal Regret in Adversarial Linear MDPs with Bandit Feedback
Towards Optimal Regret in Adversarial Linear MDPs with Bandit FeedbackInternational Conference on Learning Representations (ICLR), 2023
Haolin Liu
Chen-Yu Wei
Julian Zimmert
313
10
0
17 Oct 2023
Bypassing the Simulator: Near-Optimal Adversarial Linear Contextual
  Bandits
Bypassing the Simulator: Near-Optimal Adversarial Linear Contextual BanditsNeural Information Processing Systems (NeurIPS), 2023
Haolin Liu
Chen-Yu Wei
Julian Zimmert
291
14
0
02 Sep 2023
On Interpolating Experts and Multi-Armed Bandits
On Interpolating Experts and Multi-Armed BanditsInternational Conference on Machine Learning (ICML), 2023
Houshuang Chen
Yuchen He
Chihao Zhang
327
5
0
14 Jul 2023
Stability-penalty-adaptive follow-the-regularized-leader: Sparsity,
  game-dependency, and best-of-both-worlds
Stability-penalty-adaptive follow-the-regularized-leader: Sparsity, game-dependency, and best-of-both-worldsNeural Information Processing Systems (NeurIPS), 2023
Taira Tsuchiya
Shinji Ito
Junya Honda
326
14
0
26 May 2023
On the Minimax Regret for Online Learning with Feedback Graphs
On the Minimax Regret for Online Learning with Feedback GraphsNeural Information Processing Systems (NeurIPS), 2023
Khaled Eldowa
Emmanuel Esposito
Tommaso Cesari
Nicolò Cesa-Bianchi
248
8
0
24 May 2023
Implicitly normalized forecaster with clipping for linear and non-linear
  heavy-tailed multi-armed bandits
Implicitly normalized forecaster with clipping for linear and non-linear heavy-tailed multi-armed banditsComputational Management Science (CMS), 2023
Yuriy Dorn
Kornilov Nikita
N. Kutuzov
A. Nazin
Eduard A. Gorbunov
Alexander Gasnikov
364
5
0
11 May 2023
Accelerated Rates between Stochastic and Adversarial Online Convex Optimization
Accelerated Rates between Stochastic and Adversarial Online Convex Optimization
Sarah Sachs
Hédi Hadiji
T. Erven
Cristóbal Guzmán
326
8
0
06 Mar 2023
Best-of-Three-Worlds Linear Bandit Algorithm with Variance-Adaptive
  Regret Bounds
Best-of-Three-Worlds Linear Bandit Algorithm with Variance-Adaptive Regret BoundsAnnual Conference Computational Learning Theory (COLT), 2023
Shinji Ito
Kei Takemura
192
15
0
24 Feb 2023
1
Page 1 of 1