Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
1601.01190
Cited By
On Bayesian index policies for sequential resource allocation
6 January 2016
E. Kaufmann
Re-assign community
ArXiv
PDF
HTML
Papers citing
"On Bayesian index policies for sequential resource allocation"
35 / 35 papers shown
Title
Communication Bounds for the Distributed Experts Problem
Zhihao Jia
Qi Pang
Trung Tran
David Woodruff
Zhihao Zhang
Wenting Zheng
53
0
0
06 Jan 2025
UCB algorithms for multi-armed bandits: Precise regret and adaptive inference
Q. Han
K. Khamaru
Cun-Hui Zhang
62
3
0
09 Dec 2024
On Lai's Upper Confidence Bound in Multi-Armed Bandits
Huachen Ren
Cun-Hui Zhang
18
1
0
03 Oct 2024
Active Inference in Contextual Multi-Armed Bandits for Autonomous Robotic Exploration
Shohei Wakayama
Alberto Candela
Paul Hayne
Nisar R. Ahmed
25
0
0
07 Aug 2024
Bayesian Bandit Algorithms with Approximate Inference in Stochastic Linear Bandits
Ziyi Huang
Henry Lam
Haofeng Zhang
21
0
0
20 Jun 2024
An Experimental Design for Anytime-Valid Causal Inference on Multi-Armed Bandits
Biyonka Liang
Iavor Bojinov
30
5
0
09 Nov 2023
Simple Modification of the Upper Confidence Bound Algorithm by Generalized Weighted Averages
Nobuhito Manome
Shuji Shinohara
Ung-il Chung
11
5
0
28 Aug 2023
A General Recipe for the Analysis of Randomized Multi-Armed Bandit Algorithms
Dorian Baudry
Kazuya Suzuki
Junya Honda
16
4
0
10 Mar 2023
Optimality of Thompson Sampling with Noninformative Priors for Pareto Bandits
Jongyeong Lee
Junya Honda
Chao-Kai Chiang
Masashi Sugiyama
16
3
0
03 Feb 2023
A Combinatorial Semi-Bandit Approach to Charging Station Selection for Electric Vehicles
Niklas Åkerblom
M. Chehreghani
16
0
0
17 Jan 2023
Finite-Time Regret of Thompson Sampling Algorithms for Exponential Family Multi-Armed Bandits
Tianyuan Jin
Pan Xu
X. Xiao
Anima Anandkumar
23
12
0
07 Jun 2022
Information-Directed Selection for Top-Two Algorithms
Wei You
Chao Qin
Zihao Wang
Shuoguang Yang
20
12
0
24 May 2022
Some performance considerations when using multi-armed bandit algorithms in the presence of missing data
Xijin Chen
K. M. Lee
S. Villar
D. Robertson
26
1
0
08 May 2022
Optimal Regret Is Achievable with Bounded Approximate Inference Error: An Enhanced Bayesian Upper Confidence Bound Framework
Ziyi Huang
H. Lam
A. Meisami
Haofeng Zhang
21
4
0
31 Jan 2022
Online Learning of Energy Consumption for Navigation of Electric Vehicles
Niklas Åkerblom
Yuxin Chen
M. Chehreghani
11
12
0
03 Nov 2021
An empirical evaluation of active inference in multi-armed bandits
D. Marković
Hrvoje Stojić
Sarah Schwöbel
S. Kiebel
28
34
0
21 Jan 2021
Lifelong Learning in Multi-Armed Bandits
Matthieu Jedor
Jonathan Louëdec
Vianney Perchet
17
2
0
28 Dec 2020
MOTS: Minimax Optimal Thompson Sampling
Tianyuan Jin
Pan Xu
Jieming Shi
Xiaokui Xiao
Quanquan Gu
12
30
0
03 Mar 2020
An Online Learning Framework for Energy-Efficient Navigation of Electric Vehicles
Niklas Åkerblom
Yuxin Chen
M. Chehreghani
11
15
0
03 Mar 2020
The Unreasonable Effectiveness of Greedy Algorithms in Multi-Armed Bandit with Many Arms
Mohsen Bayati
N. Hamidi
Ramesh Johari
Khashayar Khosravi
16
28
0
24 Feb 2020
Double Explore-then-Commit: Asymptotic Optimality and Beyond
Tianyuan Jin
Pan Xu
Xiaokui Xiao
Quanquan Gu
12
24
0
21 Feb 2020
Exponential two-armed bandit problem
A. Kolnogorov
Denis Grunev
15
0
0
15 Aug 2019
Parameterized Exploration
Jesse Clifton
Lili Wu
E. Laber
19
0
0
13 Jul 2019
The Finite-Horizon Two-Armed Bandit Problem with Binary Responses: A Multidisciplinary Survey of the History, State of the Art, and Myths
P. Jacko
15
11
0
20 Jun 2019
A Note on KL-UCB+ Policy for the Stochastic Bandit
Junya Honda
8
3
0
19 Mar 2019
Adaptive Policies for Perimeter Surveillance Problems
James A. Grant
David S. Leslie
K. Glazebrook
R. Szechtman
Adam N. Letchford
13
13
0
04 Oct 2018
Profitable Bandits
Mastane Achab
Stéphan Clémençon
Aurélien Garivier
14
5
0
08 May 2018
BelMan: Bayesian Bandits on the Belief--Reward Manifold
D. Basu
Pierre Senellart
S. Bressan
11
2
0
04 May 2018
Combinatorial Multi-Armed Bandits with Filtered Feedback
James A. Grant
David S. Leslie
K. Glazebrook
R. Szechtman
19
1
0
26 May 2017
A Scale Free Algorithm for Stochastic Bandits with Bounded Kurtosis
Tor Lattimore
6
19
0
27 Mar 2017
A minimax and asymptotically optimal algorithm for stochastic bandits
Pierre Ménard
Aurélien Garivier
11
59
0
23 Feb 2017
Learning the distribution with largest mean: two bandit frameworks
E. Kaufmann
Aurélien Garivier
12
19
0
31 Jan 2017
Regret Analysis of the Anytime Optimally Confident UCB Algorithm
Tor Lattimore
12
26
0
29 Mar 2016
Simple Bayesian Algorithms for Best Arm Identification
Daniel Russo
11
273
0
26 Feb 2016
Regret Analysis of the Finite-Horizon Gittins Index Strategy for Multi-Armed Bandits
Tor Lattimore
11
46
0
18 Nov 2015
1