Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1505.04627
Cited By
Simple regret for infinitely many armed bandits
18 May 2015
Alexandra Carpentier
Michal Valko
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Simple regret for infinitely many armed bandits"
48 / 48 papers shown
Title
Sample Complexity and Representation Ability of Test-time Scaling Paradigms
Baihe Huang
Shanda Li
Tianhao Wu
Yiming Yang
Ameet Talwalkar
Kannan Ramchandran
Michael I. Jordan
Jiantao Jiao
LRM
102
0
0
05 Jun 2025
Non-Stationary Lipschitz Bandits
Nicolas Nguyen
Solenne Gaucher
Claire Vernade
39
0
0
24 May 2025
Fast and Robust: Task Sampling with Posterior and Diversity Synergies for Adaptive Decision-Makers in Randomized Environments
Yun Qu
Wenjie Wang
Yixiu Mao
Yiqin Lv
Xiangyang Ji
TTA
167
0
0
27 Apr 2025
Tracking Most Significant Shifts in Infinite-Armed Bandits
Joe Suk
Jung-hun Kim
106
1
0
31 Jan 2025
HPC Application Parameter Autotuning on Edge Devices: A Bandit Learning Approach
Abrar Hossain
Abdel-Hameed A. Badawy
M. A. Islam
Tapasya Patki
Kishwar Ahmed
74
1
0
03 Jan 2025
Exploration Unbound
Dilip Arumugam
Wanqiao Xu
Benjamin Van Roy
73
0
0
16 Jul 2024
Online Bandit Learning with Offline Preference Data for Improved RLHF
Akhil Agnihotri
Rahul Jain
Deepak Ramachandran
Zheng Wen
OffRL
217
2
0
13 Jun 2024
Practice Makes Perfect: Planning to Learn Skill Parameter Policies
Nishanth Kumar
Tom Silver
Willie McClinton
Linfeng Zhao
Stephen Proulx
Tomás Lozano-Pérez
L. Kaelbling
Jennifer Barry
114
22
0
22 Feb 2024
A Bandit Approach with Evolutionary Operators for Model Selection
Margaux Brégere Lpsm
Julie Keisler
49
1
0
07 Feb 2024
Replication-proof Bandit Mechanism Design with Bayesian Agents
Seyed A. Esmaeili
Mohammadtaghi Hajiaghayi
Suho Shin
134
1
0
28 Dec 2023
A Lipschitz Bandits Approach for Continuous Hyperparameter Optimization
Yasong Feng
Weijian Luo
Yimin Huang
Tianyu Wang
54
8
0
03 Feb 2023
Complexity Analysis of a Countable-armed Bandit Problem
Anand Kalvit
A. Zeevi
49
3
0
18 Jan 2023
On the Safety of Interpretable Machine Learning: A Maximum Deviation Approach
Dennis L. Wei
Rahul Nair
Amit Dhurandhar
Kush R. Varshney
Elizabeth M. Daly
Moninder Singh
FAtt
75
9
0
02 Nov 2022
Beyond the Best: Estimating Distribution Functionals in Infinite-Armed Bandits
Yifei Wang
Tavor Z. Baharav
Yanjun Han
Jiantao Jiao
David Tse
46
1
0
01 Nov 2022
Revisiting Simple Regret: Fast Rates for Returning a Good Arm
Yao Zhao
Connor James Stephens
Csaba Szepesvári
Kwang-Sung Jun
92
14
0
30 Oct 2022
Max-Quantile Grouped Infinite-Arm Bandits
Ivan Lau
Yan Hao Ling
Mayank Shrivastava
Jonathan Scarlett
52
1
0
04 Oct 2022
Discover Life Skills for Planning with Bandits via Observing and Learning How the World Works
Tin Lai
54
4
0
17 Jul 2022
Improving Sequential Query Recommendation with Immediate User Feedback
Shameem Puthiya Parambath
Christos Anagnostopoulos
Roderick Murray-Smith
38
1
0
12 May 2022
Non-stationary Bandits and Meta-Learning with a Small Set of Optimal Arms
Javad Azizi
T. Duong
Yasin Abbasi-Yadkori
András Gyorgy
Claire Vernade
Mohammad Ghavamzadeh
90
8
0
25 Feb 2022
Rotting Infinitely Many-armed Bandits
Jung-hun Kim
Milan Vojnović
Se-Young Yun
73
7
0
31 Jan 2022
LEGS: Learning Efficient Grasp Sets for Exploratory Grasping
Letian Fu
Michael Danielczuk
Ashwin Balakrishna
Daniel S. Brown
Jeffrey Ichnowski
Eugen Solowjow
Ken Goldberg
96
12
0
29 Nov 2021
Multi-armed Bandit Algorithm against Strategic Replication
Suho Shin
Seungjoon Lee
Jungseul Ok
57
6
0
23 Oct 2021
Bandits with Dynamic Arm-acquisition Costs
Anand Kalvit
A. Zeevi
36
3
0
23 Oct 2021
Optimal Order Simple Regret for Gaussian Process Bandits
Sattar Vakili
N. Bouziani
Sepehr Jalali
A. Bernacchia
Da-shan Shiu
99
55
0
20 Aug 2021
Periodic-GP: Learning Periodic World with Gaussian Process Bandits
Hengrui Cai
Zhihao Cen
Ling Leng
Rui Song
AI4TS
126
6
0
30 May 2021
From Finite to Countable-Armed Bandits
Anand Kalvit
A. Zeevi
76
14
0
22 May 2021
Bandits with many optimal arms
R. D. Heide
J. Cheshire
Pierre Ménard
Alexandra Carpentier
48
21
0
23 Mar 2021
Nonstochastic Bandits with Infinitely Many Experts
X. Meng
Tuhin Sarkar
M. Dahleh
OffRL
54
1
0
09 Feb 2021
Control-Data Separation and Logical Condition Propagation for Efficient Inference on Probabilistic Programs
I. Hasuo
Yuichiro Oyabu
Clovis Eberhart
Kohei Suenaga
Kenta Cho
Shin-ya Katsumata
TPM
43
3
0
05 Jan 2021
Be Greedy in Multi-Armed Bandits
Matthieu Jedor
Jonathan Louëdec
Vianney Perchet
397
8
0
04 Jan 2021
Maximal Objectives in the Multi-armed Bandit with Applications
Eren Ozbay
Vijay Kamble
61
0
0
11 Jun 2020
Sample Efficient Graph-Based Optimization with Noisy Observations
Thanh Tan Nguyen
A. Shameli
Yasin Abbasi-Yadkori
Anup B. Rao
Branislav Kveton
13
1
0
04 Jun 2020
The Unreasonable Effectiveness of Greedy Algorithms in Multi-Armed Bandit with Many Arms
Mohsen Bayati
N. Hamidi
Ramesh Johari
Khashayar Khosravi
395
29
0
24 Feb 2020
Ballooning Multi-Armed Bandits
Ganesh Ghalme
Swapnil Dhamal
Shweta Jain
Sujit Gujar
Y. Narahari
32
8
0
24 Jan 2020
A tree-based radial basis function method for noisy parallel surrogate optimization
Chenchao Shou
Matthew West
60
2
0
21 Aug 2019
The True Sample Complexity of Identifying Good Arms
Julian Katz-Samuels
Kevin Jamieson
149
42
0
15 Jun 2019
PAC Identification of Many Good Arms in Stochastic Multi-Armed Bandits
A. Chaudhuri
Shivaram Kalyanakrishnan
74
27
0
24 Jan 2019
Pure-Exploration for Infinite-Armed Bandits with General Arm Reservoirs
Maryam Aziz
Kevin Jamieson
J. Aslam
11
2
0
15 Nov 2018
Exploring
k
k
k
out of Top
ρ
ρ
ρ
Fraction of Arms in Stochastic Bandits
Wenbo Ren
Jia-Wei Liu
Ness B. Shroff
98
16
0
28 Oct 2018
Simple Regret Minimization for Contextual Bandits
A. Deshmukh
Srinagesh Sharma
J. Cutler
M. Moldwin
Clayton Scott
66
24
0
17 Oct 2018
Optimal Testing in the Experiment-rich Regime
S. Schmit
Virag Shah
Ramesh Johari
43
6
0
30 May 2018
Pure Exploration in Infinitely-Armed Bandit Models with Fixed-Confidence
Maryam Aziz
J. Anderton
E. Kaufmann
J. Aslam
23
34
0
13 Mar 2018
Max K-armed bandit: On the ExtremeHunter algorithm and beyond
Mastane Achab
Stephan Clémençon
Aurélien Garivier
Anne Sabourin
Claire Vernade
133
60
0
27 Jul 2017
On Kernelized Multi-armed Bandits
Sayak Ray Chowdhury
Aditya Gopalan
129
464
0
03 Apr 2017
On the Detection of Mixture Distributions with applications to the Most Biased Coin Problem
Kevin Jamieson
D. Haas
Ben Recht
28
2
0
25 Mar 2016
Hyperband: A Novel Bandit-Based Approach to Hyperparameter Optimization
Lisha Li
Kevin Jamieson
Giulia DeSalvo
Afshin Rostamizadeh
Ameet Talwalkar
252
2,342
0
21 Mar 2016
The Max
K
K
K
-Armed Bandit: PAC Lower Bounds and Efficient Algorithms
Y. David
N. Shimkin
27
1
0
23 Dec 2015
A Survey of Online Experiment Design with the Stochastic Multi-Armed Bandit
Giuseppe Burtini
Jason L. Loeppky
Ramon Lawrence
88
119
0
02 Oct 2015
1