ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1906.00531
  4. Cited By
Model selection for contextual bandits

Model selection for contextual bandits

3 June 2019
Dylan J. Foster
A. Krishnamurthy
Haipeng Luo
    OffRL
ArXivPDFHTML

Papers citing "Model selection for contextual bandits"

22 / 22 papers shown
Title
Budgeted Online Model Selection and Fine-Tuning via Federated Learning
Budgeted Online Model Selection and Fine-Tuning via Federated Learning
P. M. Ghari
Yanning Shen
FedML
46
1
0
19 Jan 2024
Anytime Model Selection in Linear Bandits
Anytime Model Selection in Linear Bandits
Parnian Kassraie
N. Emmenegger
Andreas Krause
Aldo Pacchiano
49
2
0
24 Jul 2023
High-dimensional Contextual Bandit Problem without Sparsity
High-dimensional Contextual Bandit Problem without Sparsity
Junpei Komiyama
Masaaki Imaizumi
34
1
0
19 Jun 2023
Active Policy Improvement from Multiple Black-box Oracles
Active Policy Improvement from Multiple Black-box Oracles
Xuefeng Liu
Takuma Yoneda
Chaoqi Wang
Matthew R. Walter
Yuxin Chen
39
9
0
17 Jun 2023
Efficient Deep Reinforcement Learning Requires Regulating Overfitting
Efficient Deep Reinforcement Learning Requires Regulating Overfitting
Qiyang Li
Aviral Kumar
Ilya Kostrikov
Sergey Levine
OffRL
29
31
0
20 Apr 2023
Estimating Optimal Policy Value in General Linear Contextual Bandits
Estimating Optimal Policy Value in General Linear Contextual Bandits
Jonathan Lee
Weihao Kong
Aldo Pacchiano
Vidya Muthukumar
Emma Brunskill
28
0
0
19 Feb 2023
Linear Bandits with Memory: from Rotting to Rising
Linear Bandits with Memory: from Rotting to Rising
Giulia Clerici
Pierre Laforgue
Nicolò Cesa-Bianchi
30
3
0
16 Feb 2023
Scalable Representation Learning in Linear Contextual Bandits with
  Constant Regret Guarantees
Scalable Representation Learning in Linear Contextual Bandits with Constant Regret Guarantees
Andrea Tirinzoni
Matteo Papini
Ahmed Touati
A. Lazaric
Matteo Pirotta
28
4
0
24 Oct 2022
Exploration in Linear Bandits with Rich Action Sets and its Implications
  for Inference
Exploration in Linear Bandits with Rich Action Sets and its Implications for Inference
Debangshu Banerjee
Avishek Ghosh
Sayak Ray Chowdhury
Aditya Gopalan
29
9
0
23 Jul 2022
Best of Both Worlds Model Selection
Best of Both Worlds Model Selection
Aldo Pacchiano
Christoph Dann
Claudio Gentile
28
10
0
29 Jun 2022
Breaking the $\sqrt{T}$ Barrier: Instance-Independent Logarithmic Regret
  in Stochastic Contextual Linear Bandits
Breaking the T\sqrt{T}T​ Barrier: Instance-Independent Logarithmic Regret in Stochastic Contextual Linear Bandits
Avishek Ghosh
Abishek Sankararaman
27
3
0
19 May 2022
Flexible and Efficient Contextual Bandits with Heterogeneous Treatment
  Effect Oracles
Flexible and Efficient Contextual Bandits with Heterogeneous Treatment Effect Oracles
Aldo G. Carranza
Sanath Kumar Krishnamurthy
Susan Athey
19
1
0
30 Mar 2022
Corralling a Larger Band of Bandits: A Case Study on Switching Regret
  for Linear Bandits
Corralling a Larger Band of Bandits: A Case Study on Switching Regret for Linear Bandits
Haipeng Luo
Mengxiao Zhang
Peng Zhao
Zhi-Hua Zhou
34
17
0
12 Feb 2022
Linear Contextual Bandits with Adversarial Corruptions
Linear Contextual Bandits with Adversarial Corruptions
Heyang Zhao
Dongruo Zhou
Quanquan Gu
AAML
28
24
0
25 Oct 2021
Near Instance Optimal Model Selection for Pure Exploration Linear
  Bandits
Near Instance Optimal Model Selection for Pure Exploration Linear Bandits
Yinglun Zhu
Julian Katz-Samuels
Robert D. Nowak
38
6
0
10 Sep 2021
Provably Efficient Representation Selection in Low-rank Markov Decision
  Processes: From Online to Offline RL
Provably Efficient Representation Selection in Low-rank Markov Decision Processes: From Online to Offline RL
Weitong Zhang
Jiafan He
Dongruo Zhou
Amy Zhang
Quanquan Gu
OffRL
22
11
0
22 Jun 2021
Leveraging Good Representations in Linear Contextual Bandits
Leveraging Good Representations in Linear Contextual Bandits
Matteo Papini
Andrea Tirinzoni
Marcello Restelli
A. Lazaric
Matteo Pirotta
30
26
0
08 Apr 2021
Minimax Regret for Stochastic Shortest Path with Adversarial Costs and
  Known Transition
Minimax Regret for Stochastic Shortest Path with Adversarial Costs and Known Transition
Liyu Chen
Haipeng Luo
Chen-Yu Wei
23
32
0
07 Dec 2020
Regret Balancing for Bandit and RL Model Selection
Regret Balancing for Bandit and RL Model Selection
Yasin Abbasi-Yadkori
Aldo Pacchiano
My Phan
21
26
0
09 Jun 2020
Rate-adaptive model selection over a collection of black-box contextual
  bandit algorithms
Rate-adaptive model selection over a collection of black-box contextual bandit algorithms
Aurélien F. Bibaut
Antoine Chambaz
Mark van der Laan
26
6
0
05 Jun 2020
Learning to Optimize under Non-Stationarity
Learning to Optimize under Non-Stationarity
Wang Chi Cheung
D. Simchi-Levi
Ruihao Zhu
36
133
0
06 Oct 2018
Estimating Learnability in the Sublinear Data Regime
Estimating Learnability in the Sublinear Data Regime
Weihao Kong
Gregory Valiant
43
29
0
04 May 2018
1