ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1505.04627
  4. Cited By
Simple regret for infinitely many armed bandits

Simple regret for infinitely many armed bandits

18 May 2015
Alexandra Carpentier
Michal Valko
ArXiv (abs)PDFHTML

Papers citing "Simple regret for infinitely many armed bandits"

48 / 48 papers shown
Title
Sample Complexity and Representation Ability of Test-time Scaling Paradigms
Sample Complexity and Representation Ability of Test-time Scaling Paradigms
Baihe Huang
Shanda Li
Tianhao Wu
Yiming Yang
Ameet Talwalkar
Kannan Ramchandran
Michael I. Jordan
Jiantao Jiao
LRM
102
0
0
05 Jun 2025
Non-Stationary Lipschitz Bandits
Non-Stationary Lipschitz Bandits
Nicolas Nguyen
Solenne Gaucher
Claire Vernade
39
0
0
24 May 2025
Fast and Robust: Task Sampling with Posterior and Diversity Synergies for Adaptive Decision-Makers in Randomized Environments
Fast and Robust: Task Sampling with Posterior and Diversity Synergies for Adaptive Decision-Makers in Randomized Environments
Yun Qu
Wenjie Wang
Yixiu Mao
Yiqin Lv
Xiangyang Ji
TTA
167
0
0
27 Apr 2025
Tracking Most Significant Shifts in Infinite-Armed Bandits
Joe Suk
Jung-hun Kim
106
1
0
31 Jan 2025
HPC Application Parameter Autotuning on Edge Devices: A Bandit Learning Approach
Abrar Hossain
Abdel-Hameed A. Badawy
M. A. Islam
Tapasya Patki
Kishwar Ahmed
74
1
0
03 Jan 2025
Exploration Unbound
Exploration Unbound
Dilip Arumugam
Wanqiao Xu
Benjamin Van Roy
73
0
0
16 Jul 2024
Online Bandit Learning with Offline Preference Data for Improved RLHF
Online Bandit Learning with Offline Preference Data for Improved RLHF
Akhil Agnihotri
Rahul Jain
Deepak Ramachandran
Zheng Wen
OffRL
217
2
0
13 Jun 2024
Practice Makes Perfect: Planning to Learn Skill Parameter Policies
Practice Makes Perfect: Planning to Learn Skill Parameter Policies
Nishanth Kumar
Tom Silver
Willie McClinton
Linfeng Zhao
Stephen Proulx
Tomás Lozano-Pérez
L. Kaelbling
Jennifer Barry
114
22
0
22 Feb 2024
A Bandit Approach with Evolutionary Operators for Model Selection
A Bandit Approach with Evolutionary Operators for Model Selection
Margaux Brégere Lpsm
Julie Keisler
49
1
0
07 Feb 2024
Replication-proof Bandit Mechanism Design with Bayesian Agents
Replication-proof Bandit Mechanism Design with Bayesian Agents
Seyed A. Esmaeili
Mohammadtaghi Hajiaghayi
Suho Shin
134
1
0
28 Dec 2023
A Lipschitz Bandits Approach for Continuous Hyperparameter Optimization
A Lipschitz Bandits Approach for Continuous Hyperparameter Optimization
Yasong Feng
Weijian Luo
Yimin Huang
Tianyu Wang
54
8
0
03 Feb 2023
Complexity Analysis of a Countable-armed Bandit Problem
Complexity Analysis of a Countable-armed Bandit Problem
Anand Kalvit
A. Zeevi
49
3
0
18 Jan 2023
On the Safety of Interpretable Machine Learning: A Maximum Deviation
  Approach
On the Safety of Interpretable Machine Learning: A Maximum Deviation Approach
Dennis L. Wei
Rahul Nair
Amit Dhurandhar
Kush R. Varshney
Elizabeth M. Daly
Moninder Singh
FAtt
75
9
0
02 Nov 2022
Beyond the Best: Estimating Distribution Functionals in Infinite-Armed
  Bandits
Beyond the Best: Estimating Distribution Functionals in Infinite-Armed Bandits
Yifei Wang
Tavor Z. Baharav
Yanjun Han
Jiantao Jiao
David Tse
46
1
0
01 Nov 2022
Revisiting Simple Regret: Fast Rates for Returning a Good Arm
Revisiting Simple Regret: Fast Rates for Returning a Good Arm
Yao Zhao
Connor James Stephens
Csaba Szepesvári
Kwang-Sung Jun
92
14
0
30 Oct 2022
Max-Quantile Grouped Infinite-Arm Bandits
Max-Quantile Grouped Infinite-Arm Bandits
Ivan Lau
Yan Hao Ling
Mayank Shrivastava
Jonathan Scarlett
52
1
0
04 Oct 2022
Discover Life Skills for Planning with Bandits via Observing and
  Learning How the World Works
Discover Life Skills for Planning with Bandits via Observing and Learning How the World Works
Tin Lai
54
4
0
17 Jul 2022
Improving Sequential Query Recommendation with Immediate User Feedback
Improving Sequential Query Recommendation with Immediate User Feedback
Shameem Puthiya Parambath
Christos Anagnostopoulos
Roderick Murray-Smith
38
1
0
12 May 2022
Non-stationary Bandits and Meta-Learning with a Small Set of Optimal
  Arms
Non-stationary Bandits and Meta-Learning with a Small Set of Optimal Arms
Javad Azizi
T. Duong
Yasin Abbasi-Yadkori
András Gyorgy
Claire Vernade
Mohammad Ghavamzadeh
90
8
0
25 Feb 2022
Rotting Infinitely Many-armed Bandits
Rotting Infinitely Many-armed Bandits
Jung-hun Kim
Milan Vojnović
Se-Young Yun
73
7
0
31 Jan 2022
LEGS: Learning Efficient Grasp Sets for Exploratory Grasping
LEGS: Learning Efficient Grasp Sets for Exploratory Grasping
Letian Fu
Michael Danielczuk
Ashwin Balakrishna
Daniel S. Brown
Jeffrey Ichnowski
Eugen Solowjow
Ken Goldberg
96
12
0
29 Nov 2021
Multi-armed Bandit Algorithm against Strategic Replication
Multi-armed Bandit Algorithm against Strategic Replication
Suho Shin
Seungjoon Lee
Jungseul Ok
57
6
0
23 Oct 2021
Bandits with Dynamic Arm-acquisition Costs
Bandits with Dynamic Arm-acquisition Costs
Anand Kalvit
A. Zeevi
36
3
0
23 Oct 2021
Optimal Order Simple Regret for Gaussian Process Bandits
Optimal Order Simple Regret for Gaussian Process Bandits
Sattar Vakili
N. Bouziani
Sepehr Jalali
A. Bernacchia
Da-shan Shiu
99
55
0
20 Aug 2021
Periodic-GP: Learning Periodic World with Gaussian Process Bandits
Periodic-GP: Learning Periodic World with Gaussian Process Bandits
Hengrui Cai
Zhihao Cen
Ling Leng
Rui Song
AI4TS
126
6
0
30 May 2021
From Finite to Countable-Armed Bandits
From Finite to Countable-Armed Bandits
Anand Kalvit
A. Zeevi
76
14
0
22 May 2021
Bandits with many optimal arms
Bandits with many optimal arms
R. D. Heide
J. Cheshire
Pierre Ménard
Alexandra Carpentier
48
21
0
23 Mar 2021
Nonstochastic Bandits with Infinitely Many Experts
Nonstochastic Bandits with Infinitely Many Experts
X. Meng
Tuhin Sarkar
M. Dahleh
OffRL
54
1
0
09 Feb 2021
Control-Data Separation and Logical Condition Propagation for Efficient
  Inference on Probabilistic Programs
Control-Data Separation and Logical Condition Propagation for Efficient Inference on Probabilistic Programs
I. Hasuo
Yuichiro Oyabu
Clovis Eberhart
Kohei Suenaga
Kenta Cho
Shin-ya Katsumata
TPM
43
3
0
05 Jan 2021
Be Greedy in Multi-Armed Bandits
Be Greedy in Multi-Armed Bandits
Matthieu Jedor
Jonathan Louëdec
Vianney Perchet
397
8
0
04 Jan 2021
Maximal Objectives in the Multi-armed Bandit with Applications
Maximal Objectives in the Multi-armed Bandit with Applications
Eren Ozbay
Vijay Kamble
61
0
0
11 Jun 2020
Sample Efficient Graph-Based Optimization with Noisy Observations
Sample Efficient Graph-Based Optimization with Noisy Observations
Thanh Tan Nguyen
A. Shameli
Yasin Abbasi-Yadkori
Anup B. Rao
Branislav Kveton
13
1
0
04 Jun 2020
The Unreasonable Effectiveness of Greedy Algorithms in Multi-Armed
  Bandit with Many Arms
The Unreasonable Effectiveness of Greedy Algorithms in Multi-Armed Bandit with Many Arms
Mohsen Bayati
N. Hamidi
Ramesh Johari
Khashayar Khosravi
395
29
0
24 Feb 2020
Ballooning Multi-Armed Bandits
Ballooning Multi-Armed Bandits
Ganesh Ghalme
Swapnil Dhamal
Shweta Jain
Sujit Gujar
Y. Narahari
32
8
0
24 Jan 2020
A tree-based radial basis function method for noisy parallel surrogate
  optimization
A tree-based radial basis function method for noisy parallel surrogate optimization
Chenchao Shou
Matthew West
60
2
0
21 Aug 2019
The True Sample Complexity of Identifying Good Arms
The True Sample Complexity of Identifying Good Arms
Julian Katz-Samuels
Kevin Jamieson
149
42
0
15 Jun 2019
PAC Identification of Many Good Arms in Stochastic Multi-Armed Bandits
PAC Identification of Many Good Arms in Stochastic Multi-Armed Bandits
A. Chaudhuri
Shivaram Kalyanakrishnan
74
27
0
24 Jan 2019
Pure-Exploration for Infinite-Armed Bandits with General Arm Reservoirs
Maryam Aziz
Kevin Jamieson
J. Aslam
11
2
0
15 Nov 2018
Exploring $k$ out of Top $ρ$ Fraction of Arms in Stochastic Bandits
Exploring kkk out of Top ρρρ Fraction of Arms in Stochastic Bandits
Wenbo Ren
Jia-Wei Liu
Ness B. Shroff
98
16
0
28 Oct 2018
Simple Regret Minimization for Contextual Bandits
Simple Regret Minimization for Contextual Bandits
A. Deshmukh
Srinagesh Sharma
J. Cutler
M. Moldwin
Clayton Scott
66
24
0
17 Oct 2018
Optimal Testing in the Experiment-rich Regime
Optimal Testing in the Experiment-rich Regime
S. Schmit
Virag Shah
Ramesh Johari
43
6
0
30 May 2018
Pure Exploration in Infinitely-Armed Bandit Models with Fixed-Confidence
Pure Exploration in Infinitely-Armed Bandit Models with Fixed-Confidence
Maryam Aziz
J. Anderton
E. Kaufmann
J. Aslam
23
34
0
13 Mar 2018
Max K-armed bandit: On the ExtremeHunter algorithm and beyond
Max K-armed bandit: On the ExtremeHunter algorithm and beyond
Mastane Achab
Stephan Clémençon
Aurélien Garivier
Anne Sabourin
Claire Vernade
133
60
0
27 Jul 2017
On Kernelized Multi-armed Bandits
On Kernelized Multi-armed Bandits
Sayak Ray Chowdhury
Aditya Gopalan
129
464
0
03 Apr 2017
On the Detection of Mixture Distributions with applications to the Most
  Biased Coin Problem
On the Detection of Mixture Distributions with applications to the Most Biased Coin Problem
Kevin Jamieson
D. Haas
Ben Recht
28
2
0
25 Mar 2016
Hyperband: A Novel Bandit-Based Approach to Hyperparameter Optimization
Hyperband: A Novel Bandit-Based Approach to Hyperparameter Optimization
Lisha Li
Kevin Jamieson
Giulia DeSalvo
Afshin Rostamizadeh
Ameet Talwalkar
252
2,342
0
21 Mar 2016
The Max $K$-Armed Bandit: PAC Lower Bounds and Efficient Algorithms
The Max KKK-Armed Bandit: PAC Lower Bounds and Efficient Algorithms
Y. David
N. Shimkin
27
1
0
23 Dec 2015
A Survey of Online Experiment Design with the Stochastic Multi-Armed
  Bandit
A Survey of Online Experiment Design with the Stochastic Multi-Armed Bandit
Giuseppe Burtini
Jason L. Loeppky
Ramon Lawrence
88
119
0
02 Oct 2015
1