ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2007.00953
  4. Cited By
Gamification of Pure Exploration for Linear Bandits

Gamification of Pure Exploration for Linear Bandits

2 July 2020
Rémy Degenne
Pierre Ménard
Xuedong Shang
Michal Valko
ArXiv (abs)PDFHTML

Papers citing "Gamification of Pure Exploration for Linear Bandits"

50 / 61 papers shown
Title
Experimental Design for Semiparametric Bandits
Experimental Design for Semiparametric Bandits
Seok-Jin Kim
Gi-Soo Kim
Min-hwan Oh
21
0
0
16 Jun 2025
Sample Efficient Demonstration Selection for In-Context Learning
Kiran Purohit
Venktesh V
Sourangshu Bhattacharya
Avishek Anand
41
0
0
10 Jun 2025
Adapting to Heterophilic Graph Data with Structure-Guided Neighbor Discovery
Victor M. Tenorio
Madeline Navarro
Samuel Rey
Santiago Segarra
Antonio G. Marques
10
0
0
10 Jun 2025
Pure Exploration with Infinite Answers
Pure Exploration with Infinite Answers
Riccardo Poiani
Martino Bernasconi
A. Celli
32
0
0
28 May 2025
On the Problem of Best Arm Retention
On the Problem of Best Arm Retention
Houshuang Chen
Yuchen He
Chihao Zhang
81
0
0
16 Apr 2025
Sequential Learning of the Pareto Front for Multi-objective Bandits
Sequential Learning of the Pareto Front for Multi-objective Bandits
Elise Crépon
Aurélien Garivier
Wouter M. Koolen
87
5
0
29 Jan 2025
Enhancing Preference-based Linear Bandits via Human Response Time
Enhancing Preference-based Linear Bandits via Human Response Time
Shen Li
Yuyang Zhang
Tongzheng Ren
Claire Liang
Na Li
J. Shah
183
1
0
03 Jan 2025
Near Optimal Pure Exploration in Logistic Bandits
Near Optimal Pure Exploration in Logistic Bandits
Eduardo Ochoa Rivera
Ambuj Tewari
94
0
0
28 Oct 2024
Optimal Design for Reward Modeling in RLHF
Optimal Design for Reward Modeling in RLHF
Antoine Scheid
Etienne Boursier
Alain Durmus
Michael I. Jordan
Pierre Ménard
Eric Moulines
Michal Valko
OffRL
148
9
0
22 Oct 2024
Linear Submodular Maximization with Bandit Feedback
Linear Submodular Maximization with Bandit Feedback
Wenjing Chen
Victoria G. Crawford
20
0
0
02 Jul 2024
Pretraining Decision Transformers with Reward Prediction for In-Context Multi-task Structured Bandit Learning
Pretraining Decision Transformers with Reward Prediction for In-Context Multi-task Structured Bandit Learning
Subhojyoti Mukherjee
Josiah P. Hanna
Qiaomin Xie
Robert Nowak
248
2
0
07 Jun 2024
Regret Minimization via Saddle Point Optimization
Regret Minimization via Saddle Point Optimization
Johannes Kirschner
Seyed Alireza Bakhtiari
Kushagra Chandak
Volodymyr Tkachuk
Csaba Szepesvári
65
1
0
15 Mar 2024
LinearAPT: An Adaptive Algorithm for the Fixed-Budget Thresholding
  Linear Bandit Problem
LinearAPT: An Adaptive Algorithm for the Fixed-Budget Thresholding Linear Bandit Problem
Yun-Ang Wu
Yun-Da Tsai
Shou-De Lin
102
1
0
10 Mar 2024
Optimal Thresholding Linear Bandit
Optimal Thresholding Linear Bandit
Eduardo Ochoa Rivera
Ambuj Tewari
57
0
0
11 Feb 2024
Differentially Private High Dimensional Bandits
Differentially Private High Dimensional Bandits
Apurv Shukla
59
0
0
06 Feb 2024
Adaptive Experiment Design with Synthetic Controls
Adaptive Experiment Design with Synthetic Controls
Alihan Huyuk
Zhaozhi Qian
M. Schaar
17
2
0
30 Jan 2024
Experiment Planning with Function Approximation
Experiment Planning with Function Approximation
Aldo Pacchiano
Jonathan Lee
Emma Brunskill
OffRL
70
4
0
10 Jan 2024
Robust Best-arm Identification in Linear Bandits
Robust Best-arm Identification in Linear Bandits
Wei Wang
Sattar Vakili
Ilija Bogunovic
67
0
0
08 Nov 2023
Optimal Batched Best Arm Identification
Optimal Batched Best Arm Identification
Tianyuan Jin
Yu Yang
Jing Tang
Xiaokui Xiao
Pan Xu
116
3
0
21 Oct 2023
Pure Exploration in Asynchronous Federated Bandits
Pure Exploration in Asynchronous Federated Bandits
Zichen Wang
Chuanhao Li
Chenyu Song
Lianghui Wang
Quanquan Gu
Huazheng Wang
FedML
69
3
0
17 Oct 2023
Optimal Exploration is no harder than Thompson Sampling
Optimal Exploration is no harder than Thompson Sampling
Zhaoqi Li
Kevin Jamieson
Lalit P. Jain
69
3
0
09 Oct 2023
Experimental Designs for Heteroskedastic Variance
Experimental Designs for Heteroskedastic Variance
Justin Weltz
Tanner Fiez
Alex Volfovsky
Eric B. Laber
Blake Mason
Houssam Nassif
Lalit P. Jain
84
5
0
06 Oct 2023
Price of Safety in Linear Best Arm Identification
Price of Safety in Linear Best Arm Identification
Xuedong Shang
Igor Colin
M. Barlier
Hamza Cherkaoui
LLMSV
58
5
0
15 Sep 2023
Pure Exploration under Mediators' Feedback
Pure Exploration under Mediators' Feedback
Riccardo Poiani
Alberto Maria Metelli
Marcello Restelli
50
1
0
29 Aug 2023
Certified Multi-Fidelity Zeroth-Order Optimization
Certified Multi-Fidelity Zeroth-Order Optimization
Étienne de Montbrun
Sébastien Gerchinovitz
86
1
0
02 Aug 2023
A/B Testing and Best-arm Identification for Linear Bandits with
  Robustness to Non-stationarity
A/B Testing and Best-arm Identification for Linear Bandits with Robustness to Non-stationarity
Zhihan Xiong
Romain Camilleri
Maryam Fazel
Lalit P. Jain
Kevin Jamieson
137
1
0
27 Jul 2023
Pure Exploration in Bandits with Linear Constraints
Pure Exploration in Bandits with Linear Constraints
Emil Carlsson
Debabrota Basu
Fredrik D. Johansson
Devdatt Dubhashi
73
4
0
22 Jun 2023
Cooperative Thresholded Lasso for Sparse Linear Bandit
Cooperative Thresholded Lasso for Sparse Linear Bandit
Haniyeh Barghi
Xiaotong Cheng
S. Maghsudi
80
0
0
30 May 2023
Multi-task Representation Learning for Pure Exploration in Linear
  Bandits
Multi-task Representation Learning for Pure Exploration in Linear Bandits
Yihan Du
Longbo Huang
Wen Sun
99
4
0
09 Feb 2023
An Asymptotically Optimal Algorithm for the Convex Hull Membership
  Problem
An Asymptotically Optimal Algorithm for the Convex Hull Membership Problem
Gang Qiao
Ambuj Tewari
53
0
0
03 Feb 2023
Best Arm Identification in Stochastic Bandits: Beyond $β-$optimality
Best Arm Identification in Stochastic Bandits: Beyond β−β-β−optimality
Arpan Mukherjee
A. Tajer
64
3
0
10 Jan 2023
Scalable Representation Learning in Linear Contextual Bandits with
  Constant Regret Guarantees
Scalable Representation Learning in Linear Contextual Bandits with Constant Regret Guarantees
Andrea Tirinzoni
Matteo Papini
Ahmed Touati
A. Lazaric
Matteo Pirotta
68
4
0
24 Oct 2022
SPRT-based Efficient Best Arm Identification in Stochastic Bandits
SPRT-based Efficient Best Arm Identification in Stochastic Bandits
Arpan Mukherjee
A. Tajer
66
6
0
22 Jul 2022
Instance-optimal PAC Algorithms for Contextual Bandits
Instance-optimal PAC Algorithms for Contextual Bandits
Zhao Li
Lillian J. Ratliff
Houssam Nassif
Kevin Jamieson
Lalit P. Jain
95
18
0
05 Jul 2022
Active Learning with Safety Constraints
Active Learning with Safety Constraints
Romain Camilleri
Andrew Wagenmaker
Jamie Morgenstern
Lalit P. Jain
Kevin Jamieson
66
14
0
22 Jun 2022
Choosing Answers in $\varepsilon$-Best-Answer Identification for Linear
  Bandits
Choosing Answers in ε\varepsilonε-Best-Answer Identification for Linear Bandits
Marc Jourdan
Rémy Degenne
42
1
0
09 Jun 2022
On Elimination Strategies for Bandit Fixed-Confidence Identification
On Elimination Strategies for Bandit Fixed-Confidence Identification
Andrea Tirinzoni
Rémy Degenne
88
7
0
22 May 2022
Instance-Dependent Regret Analysis of Kernelized Bandits
Instance-Dependent Regret Analysis of Kernelized Bandits
S. Shekhar
T. Javidi
62
4
0
12 Mar 2022
Nearly Optimal Algorithms for Level Set Estimation
Nearly Optimal Algorithms for Level Set Estimation
Blake Mason
Romain Camilleri
Subhojyoti Mukherjee
Kevin Jamieson
Robert D. Nowak
Lalit P. Jain
77
23
0
02 Nov 2021
Dealing With Misspecification In Fixed-Confidence Linear Top-m
  Identification
Dealing With Misspecification In Fixed-Confidence Linear Top-m Identification
Clémence Réda
Andrea Tirinzoni
Rémy Degenne
64
10
0
02 Nov 2021
Vector Optimization with Stochastic Bandit Feedback
Vector Optimization with Stochastic Bandit Feedback
Shiao Liu
Jian Huang
68
10
0
23 Oct 2021
Near Instance Optimal Model Selection for Pure Exploration Linear
  Bandits
Near Instance Optimal Model Selection for Pure Exploration Linear Bandits
Yinglun Zhu
Julian Katz-Samuels
Robert D. Nowak
62
7
0
10 Sep 2021
Design of Experiments for Stochastic Contextual Linear Bandits
Design of Experiments for Stochastic Contextual Linear Bandits
Andrea Zanette
Kefan Dong
Jonathan Lee
Emma Brunskill
OffRL
75
18
0
21 Jul 2021
Pure Exploration in Kernel and Neural Bandits
Pure Exploration in Kernel and Neural Bandits
Yinglun Zhu
Dongruo Zhou
Ruoxi Jiang
Quanquan Gu
Rebecca Willett
Robert D. Nowak
67
16
0
22 Jun 2021
Fixed-Budget Best-Arm Identification in Structured Bandits
Fixed-Budget Best-Arm Identification in Structured Bandits
Javad Azizi
Branislav Kveton
Mohammad Ghavamzadeh
167
26
0
09 Jun 2021
Minimax Optimal Fixed-Budget Best Arm Identification in Linear Bandits
Minimax Optimal Fixed-Budget Best Arm Identification in Linear Bandits
Junwen Yang
Vincent Y. F. Tan
67
26
0
27 May 2021
High-Dimensional Experimental Design and Kernel Bandits
High-Dimensional Experimental Design and Kernel Bandits
Romain Camilleri
Julian Katz-Samuels
Kevin Jamieson
83
57
0
12 May 2021
Pure Exploration with Structured Preference Feedback
Pure Exploration with Structured Preference Feedback
Shubham Gupta
Aadirupa Saha
S. Katariya
86
0
0
12 Apr 2021
Leveraging Good Representations in Linear Contextual Bandits
Leveraging Good Representations in Linear Contextual Bandits
Matteo Papini
Andrea Tirinzoni
Marcello Restelli
A. Lazaric
Matteo Pirotta
73
27
0
08 Apr 2021
Top-m identification for linear bandits
Top-m identification for linear bandits
Clémence Réda
E. Kaufmann
A. Delahaye-Duriez
74
15
0
18 Mar 2021
12
Next