Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1906.00531
Cited By
Model selection for contextual bandits
3 June 2019
Dylan J. Foster
A. Krishnamurthy
Haipeng Luo
OffRL
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Model selection for contextual bandits"
22 / 22 papers shown
Title
Budgeted Online Model Selection and Fine-Tuning via Federated Learning
P. M. Ghari
Yanning Shen
FedML
46
1
0
19 Jan 2024
Anytime Model Selection in Linear Bandits
Parnian Kassraie
N. Emmenegger
Andreas Krause
Aldo Pacchiano
49
2
0
24 Jul 2023
High-dimensional Contextual Bandit Problem without Sparsity
Junpei Komiyama
Masaaki Imaizumi
34
1
0
19 Jun 2023
Active Policy Improvement from Multiple Black-box Oracles
Xuefeng Liu
Takuma Yoneda
Chaoqi Wang
Matthew R. Walter
Yuxin Chen
39
9
0
17 Jun 2023
Efficient Deep Reinforcement Learning Requires Regulating Overfitting
Qiyang Li
Aviral Kumar
Ilya Kostrikov
Sergey Levine
OffRL
29
31
0
20 Apr 2023
Estimating Optimal Policy Value in General Linear Contextual Bandits
Jonathan Lee
Weihao Kong
Aldo Pacchiano
Vidya Muthukumar
Emma Brunskill
28
0
0
19 Feb 2023
Linear Bandits with Memory: from Rotting to Rising
Giulia Clerici
Pierre Laforgue
Nicolò Cesa-Bianchi
30
3
0
16 Feb 2023
Scalable Representation Learning in Linear Contextual Bandits with Constant Regret Guarantees
Andrea Tirinzoni
Matteo Papini
Ahmed Touati
A. Lazaric
Matteo Pirotta
28
4
0
24 Oct 2022
Exploration in Linear Bandits with Rich Action Sets and its Implications for Inference
Debangshu Banerjee
Avishek Ghosh
Sayak Ray Chowdhury
Aditya Gopalan
29
9
0
23 Jul 2022
Best of Both Worlds Model Selection
Aldo Pacchiano
Christoph Dann
Claudio Gentile
28
10
0
29 Jun 2022
Breaking the
T
\sqrt{T}
T
Barrier: Instance-Independent Logarithmic Regret in Stochastic Contextual Linear Bandits
Avishek Ghosh
Abishek Sankararaman
27
3
0
19 May 2022
Flexible and Efficient Contextual Bandits with Heterogeneous Treatment Effect Oracles
Aldo G. Carranza
Sanath Kumar Krishnamurthy
Susan Athey
19
1
0
30 Mar 2022
Corralling a Larger Band of Bandits: A Case Study on Switching Regret for Linear Bandits
Haipeng Luo
Mengxiao Zhang
Peng Zhao
Zhi-Hua Zhou
34
17
0
12 Feb 2022
Linear Contextual Bandits with Adversarial Corruptions
Heyang Zhao
Dongruo Zhou
Quanquan Gu
AAML
28
24
0
25 Oct 2021
Near Instance Optimal Model Selection for Pure Exploration Linear Bandits
Yinglun Zhu
Julian Katz-Samuels
Robert D. Nowak
38
6
0
10 Sep 2021
Provably Efficient Representation Selection in Low-rank Markov Decision Processes: From Online to Offline RL
Weitong Zhang
Jiafan He
Dongruo Zhou
Amy Zhang
Quanquan Gu
OffRL
22
11
0
22 Jun 2021
Leveraging Good Representations in Linear Contextual Bandits
Matteo Papini
Andrea Tirinzoni
Marcello Restelli
A. Lazaric
Matteo Pirotta
30
26
0
08 Apr 2021
Minimax Regret for Stochastic Shortest Path with Adversarial Costs and Known Transition
Liyu Chen
Haipeng Luo
Chen-Yu Wei
23
32
0
07 Dec 2020
Regret Balancing for Bandit and RL Model Selection
Yasin Abbasi-Yadkori
Aldo Pacchiano
My Phan
21
26
0
09 Jun 2020
Rate-adaptive model selection over a collection of black-box contextual bandit algorithms
Aurélien F. Bibaut
Antoine Chambaz
Mark van der Laan
26
6
0
05 Jun 2020
Learning to Optimize under Non-Stationarity
Wang Chi Cheung
D. Simchi-Levi
Ruihao Zhu
36
133
0
06 Oct 2018
Estimating Learnability in the Sublinear Data Regime
Weihao Kong
Gregory Valiant
43
29
0
04 May 2018
1