ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2003.01704
  4. Cited By
Model Selection in Contextual Stochastic Bandit Problems

Model Selection in Contextual Stochastic Bandit Problems

3 March 2020
Aldo Pacchiano
My Phan
Yasin Abbasi-Yadkori
Anup B. Rao
Julian Zimmert
Tor Lattimore
Csaba Szepesvári
ArXivPDFHTML

Papers citing "Model Selection in Contextual Stochastic Bandit Problems"

23 / 23 papers shown
Title
Causal Bandits: The Pareto Optimal Frontier of Adaptivity, a Reduction
  to Linear Bandits, and Limitations around Unknown Marginals
Causal Bandits: The Pareto Optimal Frontier of Adaptivity, a Reduction to Linear Bandits, and Limitations around Unknown Marginals
Ziyi Liu
Idan Attias
Daniel M. Roy
CML
34
0
0
01 Jul 2024
Budgeted Online Model Selection and Fine-Tuning via Federated Learning
Budgeted Online Model Selection and Fine-Tuning via Federated Learning
P. M. Ghari
Yanning Shen
FedML
50
1
0
19 Jan 2024
Anytime Model Selection in Linear Bandits
Anytime Model Selection in Linear Bandits
Parnian Kassraie
N. Emmenegger
Andreas Krause
Aldo Pacchiano
49
2
0
24 Jul 2023
Active Policy Improvement from Multiple Black-box Oracles
Active Policy Improvement from Multiple Black-box Oracles
Xuefeng Liu
Takuma Yoneda
Chaoqi Wang
Matthew R. Walter
Yuxin Chen
39
9
0
17 Jun 2023
Robust Lipschitz Bandits to Adversarial Corruptions
Robust Lipschitz Bandits to Adversarial Corruptions
Yue Kang
Cho-Jui Hsieh
T. C. Lee
AAML
30
8
0
29 May 2023
Estimating Optimal Policy Value in General Linear Contextual Bandits
Estimating Optimal Policy Value in General Linear Contextual Bandits
Jonathan Lee
Weihao Kong
Aldo Pacchiano
Vidya Muthukumar
Emma Brunskill
28
0
0
19 Feb 2023
Linear Bandits with Memory: from Rotting to Rising
Linear Bandits with Memory: from Rotting to Rising
Giulia Clerici
Pierre Laforgue
Nicolò Cesa-Bianchi
33
3
0
16 Feb 2023
Contexts can be Cheap: Solving Stochastic Contextual Bandits with Linear
  Bandit Algorithms
Contexts can be Cheap: Solving Stochastic Contextual Bandits with Linear Bandit Algorithms
Osama A. Hanna
Lin F. Yang
Christina Fragouli
27
11
0
08 Nov 2022
Scalable Representation Learning in Linear Contextual Bandits with
  Constant Regret Guarantees
Scalable Representation Learning in Linear Contextual Bandits with Constant Regret Guarantees
Andrea Tirinzoni
Matteo Papini
Ahmed Touati
A. Lazaric
Matteo Pirotta
30
4
0
24 Oct 2022
Best of Both Worlds Model Selection
Best of Both Worlds Model Selection
Aldo Pacchiano
Christoph Dann
Claudio Gentile
34
10
0
29 Jun 2022
Adversarial Bandits against Arbitrary Strategies
Adversarial Bandits against Arbitrary Strategies
Jung-hun Kim
Se-Young Yun
49
0
0
30 May 2022
Corralling a Larger Band of Bandits: A Case Study on Switching Regret
  for Linear Bandits
Corralling a Larger Band of Bandits: A Case Study on Switching Regret for Linear Bandits
Haipeng Luo
Mengxiao Zhang
Peng Zhao
Zhi-Hua Zhou
34
17
0
12 Feb 2022
Misspecified Gaussian Process Bandit Optimization
Misspecified Gaussian Process Bandit Optimization
Ilija Bogunovic
Andreas Krause
57
42
0
09 Nov 2021
Dealing With Misspecification In Fixed-Confidence Linear Top-m
  Identification
Dealing With Misspecification In Fixed-Confidence Linear Top-m Identification
Clémence Réda
Andrea Tirinzoni
Rémy Degenne
31
9
0
02 Nov 2021
Linear Contextual Bandits with Adversarial Corruptions
Linear Contextual Bandits with Adversarial Corruptions
Heyang Zhao
Dongruo Zhou
Quanquan Gu
AAML
36
24
0
25 Oct 2021
Near Instance Optimal Model Selection for Pure Exploration Linear
  Bandits
Near Instance Optimal Model Selection for Pure Exploration Linear Bandits
Yinglun Zhu
Julian Katz-Samuels
Robert D. Nowak
38
6
0
10 Sep 2021
Provably Efficient Representation Selection in Low-rank Markov Decision
  Processes: From Online to Offline RL
Provably Efficient Representation Selection in Low-rank Markov Decision Processes: From Online to Offline RL
Weitong Zhang
Jiafan He
Dongruo Zhou
Amy Zhang
Quanquan Gu
OffRL
22
11
0
22 Jun 2021
Neural Active Learning with Performance Guarantees
Neural Active Learning with Performance Guarantees
Pranjal Awasthi
Christoph Dann
Claudio Gentile
Ayush Sekhari
Zhilei Wang
32
22
0
06 Jun 2021
Human-AI Collaboration with Bandit Feedback
Human-AI Collaboration with Bandit Feedback
Ruijiang Gao
M. Saar-Tsechansky
Maria De-Arteaga
Ligong Han
Min Kyung Lee
Matthew Lease
54
49
0
22 May 2021
Leveraging Good Representations in Linear Contextual Bandits
Leveraging Good Representations in Linear Contextual Bandits
Matteo Papini
Andrea Tirinzoni
Marcello Restelli
A. Lazaric
Matteo Pirotta
33
26
0
08 Apr 2021
Policy Optimization as Online Learning with Mediator Feedback
Policy Optimization as Online Learning with Mediator Feedback
Alberto Maria Metelli
Matteo Papini
P. DÓro
Marcello Restelli
OffRL
27
10
0
15 Dec 2020
Regret Balancing for Bandit and RL Model Selection
Regret Balancing for Bandit and RL Model Selection
Yasin Abbasi-Yadkori
Aldo Pacchiano
My Phan
21
26
0
09 Jun 2020
Rate-adaptive model selection over a collection of black-box contextual
  bandit algorithms
Rate-adaptive model selection over a collection of black-box contextual bandit algorithms
Aurélien F. Bibaut
Antoine Chambaz
Mark van der Laan
32
6
0
05 Jun 2020
1