ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2105.11126
17
1

Cascading Bandit under Differential Privacy

24 May 2021
Kun Wang
Jing Dong
Baoxiang Wang
Shuai Li
Shuo Shao
ArXivPDFHTML
Abstract

This paper studies \emph{differential privacy (DP)} and \emph{local differential privacy (LDP)} in cascading bandits. Under DP, we propose an algorithm which guarantees ϵ\epsilonϵ-indistinguishability and a regret of O((log⁡Tϵ)1+ξ)\mathcal{O}((\frac{\log T}{\epsilon})^{1+\xi})O((ϵlogT​)1+ξ) for an arbitrarily small ξ\xiξ. This is a significant improvement from the previous work of O(log⁡3Tϵ)\mathcal{O}(\frac{\log^3 T}{\epsilon})O(ϵlog3T​) regret. Under (ϵ\epsilonϵ,δ\deltaδ)-LDP, we relax the K2K^2K2 dependence through the tradeoff between privacy budget ϵ\epsilonϵ and error probability δ\deltaδ, and obtain a regret of O(Klog⁡(1/δ)log⁡Tϵ2)\mathcal{O}(\frac{K\log (1/\delta) \log T}{\epsilon^2})O(ϵ2Klog(1/δ)logT​), where KKK is the size of the arm subset. This result holds for both Gaussian mechanism and Laplace mechanism by analyses on the composition. Our results extend to combinatorial semi-bandit. We show respective lower bounds for DP and LDP cascading bandits. Extensive experiments corroborate our theoretic findings.

View on arXiv
Comments on this paper