ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1905.12624
  4. Cited By
Top-k Combinatorial Bandits with Full-Bandit Feedback
v1v2 (latest)

Top-k Combinatorial Bandits with Full-Bandit Feedback

28 May 2019
Idan Rejwan
Yishay Mansour
ArXiv (abs)PDFHTML

Papers citing "Top-k Combinatorial Bandits with Full-Bandit Feedback"

35 / 35 papers shown
Title
Bi-Criteria Optimization for Combinatorial Bandits: Sublinear Regret and Constraint Violation under Bandit Feedback
Bi-Criteria Optimization for Combinatorial Bandits: Sublinear Regret and Constraint Violation under Bandit Feedback
Vaneet Aggarwal
Shweta Jain
Subham Pokhriyal
Christopher J. Quinn
519
0
0
15 Mar 2025
Bandit and Delayed Feedback in Online Structured Prediction
Bandit and Delayed Feedback in Online Structured Prediction
Yuki Shibukawa
Taira Tsuchiya
Shinsaku Sakaue
Kenji Yamanishi
OffRL
102
0
0
26 Feb 2025
On the Low-Complexity of Fair Learning for Combinatorial Multi-Armed Bandit
On the Low-Complexity of Fair Learning for Combinatorial Multi-Armed Bandit
Xiaoyi Wu
Bo Ji
Bin Li
FaML
107
0
0
01 Jan 2025
Effective Off-Policy Evaluation and Learning in Contextual Combinatorial
  Bandits
Effective Off-Policy Evaluation and Learning in Contextual Combinatorial Bandits
Tatsuhiro Shimizu
Koichi Tanaka
Ren Kishimoto
Haruka Kiyohara
Masahiro Nomura
Yuta Saito
CMLOffRL
88
1
0
20 Aug 2024
Combining Diverse Information for Coordinated Action: Stochastic Bandit
  Algorithms for Heterogeneous Agents
Combining Diverse Information for Coordinated Action: Stochastic Bandit Algorithms for Heterogeneous Agents
Lucia Gordon
Esther Rolf
Milind Tambe
45
1
0
06 Aug 2024
A Contextual Combinatorial Bandit Approach to Negotiation
A Contextual Combinatorial Bandit Approach to Negotiation
Yexin Li
Zhancun Mu
Siyuan Qi
54
2
0
30 Jun 2024
No-Regret M${}^{\natural}$-Concave Function Maximization: Stochastic
  Bandit Algorithms and NP-Hardness of Adversarial Full-Information Setting
No-Regret M♮{}^{\natural}♮-Concave Function Maximization: Stochastic Bandit Algorithms and NP-Hardness of Adversarial Full-Information Setting
Taihei Oki
Shinsaku Sakaue
83
0
0
21 May 2024
Nearly Minimax Optimal Regret for Multinomial Logistic Bandit
Nearly Minimax Optimal Regret for Multinomial Logistic Bandit
Joongkyu Lee
Min-hwan Oh
93
7
0
16 May 2024
Combinatorial Stochastic-Greedy Bandit
Combinatorial Stochastic-Greedy Bandit
Fares Fourati
Christopher J. Quinn
Mohamed-Slim Alouini
Vaneet Aggarwal
78
9
0
13 Dec 2023
Master-slave Deep Architecture for Top-K Multi-armed Bandits with
  Non-linear Bandit Feedback and Diversity Constraints
Master-slave Deep Architecture for Top-K Multi-armed Bandits with Non-linear Bandit Feedback and Diversity Constraints
Han Huang
Li Shen
Deheng Ye
Wei Liu
29
0
0
24 Aug 2023
Combinatorial Bandits for Maximum Value Reward Function under Max
  Value-Index Feedback
Combinatorial Bandits for Maximum Value Reward Function under Max Value-Index Feedback
Yiliu Wang
Wei Chen
Milan Vojnović
21
3
0
25 May 2023
Stochastic Submodular Bandits with Delayed Composite Anonymous Bandit Feedback
Stochastic Submodular Bandits with Delayed Composite Anonymous Bandit Feedback
M. Pedramfar
Vaneet Aggarwal
79
2
0
23 Mar 2023
Neural Bandits for Data Mining: Searching for Dangerous Polypharmacy
Neural Bandits for Data Mining: Searching for Dangerous Polypharmacy
Alexandre Larouche
Audrey Durand
Richard Khoury
C. Sirois
16
0
0
10 Dec 2022
Top-k data selection via distributed sample quantile inference
Top-k data selection via distributed sample quantile inference
Xu Zhang
M. Vasconcelos
57
1
0
01 Dec 2022
Mixed-Effect Thompson Sampling
Mixed-Effect Thompson Sampling
Imad Aouali
Branislav Kveton
S. Katariya
OffRL
90
12
0
30 May 2022
Gaussian Process Bandits with Aggregated Feedback
Gaussian Process Bandits with Aggregated Feedback
Mengyan Zhang
Russell Tsuchida
Cheng Soon Ong
53
6
0
24 Dec 2021
Scaling Blockchains: Can Committee-Based Consensus Help?
Scaling Blockchains: Can Committee-Based Consensus Help?
Alon Benhaim
B. Falk
Gerry Tsoukalas
48
14
0
16 Oct 2021
Pure Exploration and Regret Minimization in Matching Bandits
Pure Exploration and Regret Minimization in Matching Bandits
Flore Sentenac
Jialin Yi
Clément Calauzènes
Vianney Perchet
Milan Vojnović
28
6
0
31 Jul 2021
Dueling Bandits with Team Comparisons
Dueling Bandits with Team Comparisons
Lee Cohen
Ulrike Schmidt-Kraepelin
Yishay Mansour
22
1
0
06 Jul 2021
Pure Exploration with Structured Preference Feedback
Pure Exploration with Structured Preference Feedback
Shubham Gupta
Aadirupa Saha
S. Katariya
70
0
0
12 Apr 2021
Efficient Optimal Selection for Composited Advertising Creatives with
  Tree Structure
Efficient Optimal Selection for Composited Advertising Creatives with Tree Structure
Jin Chen
T. Ge
Gangwei Jiang
Qing Cui
Defu Lian
Kai Zheng
42
5
0
02 Mar 2021
Top-$k$ eXtreme Contextual Bandits with Arm Hierarchy
Top-kkk eXtreme Contextual Bandits with Arm Hierarchy
Rajat Sen
Alexander Rakhlin
Lexing Ying
Rahul Kidambi
Dean Phillips Foster
Daniel N. Hill
Inderjit Dhillon
116
10
0
15 Feb 2021
Adversarial Combinatorial Bandits with General Non-linear Reward
  Functions
Adversarial Combinatorial Bandits with General Non-linear Reward Functions
Xi Chen
Yanjun Han
Yining Wang
66
17
0
05 Jan 2021
Combinatorial Pure Exploration with Full-bandit Feedback and Beyond:
  Solving Combinatorial Optimization under Uncertainty with Limited Observation
Combinatorial Pure Exploration with Full-bandit Feedback and Beyond: Solving Combinatorial Optimization under Uncertainty with Limited Observation
Yuko Kuroki
Junya Honda
Masashi Sugiyama
OffRL
46
1
0
31 Dec 2020
Fully Gap-Dependent Bounds for Multinomial Logit Bandit
Fully Gap-Dependent Bounds for Multinomial Logit Bandit
Jiaqi Yang
106
2
0
19 Nov 2020
DART: aDaptive Accept RejecT for non-linear top-K subset identification
DART: aDaptive Accept RejecT for non-linear top-K subset identification
Mridul Agarwal
Vaneet Aggarwal
Christopher J. Quinn
A. Umrawal
14
4
0
16 Nov 2020
Probabilistic Sequential Shrinking: A Best Arm Identification Algorithm
  for Stochastic Bandits with Corruptions
Probabilistic Sequential Shrinking: A Best Arm Identification Algorithm for Stochastic Bandits with Corruptions
Zixin Zhong
Wang Chi Cheung
Vincent Y. F. Tan
AAML
54
13
0
15 Oct 2020
Online Dense Subgraph Discovery via Blurred-Graph Feedback
Online Dense Subgraph Discovery via Blurred-Graph Feedback
Yuko Kuroki
Atsushi Miyauchi
Junya Honda
Masashi Sugiyama
66
16
0
24 Jun 2020
Preference-based Reinforcement Learning with Finite-Time Guarantees
Preference-based Reinforcement Learning with Finite-Time Guarantees
Yichong Xu
Ruosong Wang
Lin F. Yang
Aarti Singh
A. Dubrawski
106
60
0
16 Jun 2020
Combinatorial Pure Exploration with Full-Bandit or Partial Linear
  Feedback
Combinatorial Pure Exploration with Full-Bandit or Partial Linear Feedback
Yihan Du
Yuko Kuroki
Wei Chen
OffRL
24
4
0
14 Jun 2020
Statistically Efficient, Polynomial Time Algorithms for Combinatorial
  Semi Bandits
Statistically Efficient, Polynomial Time Algorithms for Combinatorial Semi Bandits
Thibaut Cuvelier
Richard Combes
É. Gourdin
44
21
0
17 Feb 2020
Tight Lower Bounds for Combinatorial Multi-Armed Bandits
Tight Lower Bounds for Combinatorial Multi-Armed Bandits
Nadav Merlis
Shie Mannor
42
19
0
13 Feb 2020
Best Arm Identification for Cascading Bandits in the Fixed Confidence
  Setting
Best Arm Identification for Cascading Bandits in the Fixed Confidence Setting
Zixin Zhong
Wang Chi Cheung
Vincent Y. F. Tan
63
8
0
23 Jan 2020
Thompson Sampling for Combinatorial Network Optimization in Unknown
  Environments
Thompson Sampling for Combinatorial Network Optimization in Unknown Environments
Alihan Huyuk
Cem Tekin
69
16
0
07 Jul 2019
Stochastic Top-$K$ Subset Bandits with Linear Space and Non-Linear
  Feedback
Stochastic Top-KKK Subset Bandits with Linear Space and Non-Linear Feedback
Mridul Agarwal
Vaneet Aggarwal
Christopher J. Quinn
A. Umrawal
35
10
0
29 Nov 2018
1