ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2305.18442
39
3

Improved Projection-free Online Continuous Submodular Maximization

29 May 2023
Yucheng Liao
Yuanyu Wan
Chang Yao
Mingli Song
ArXivPDFHTML
Abstract

We investigate the problem of online learning with monotone and continuous DR-submodular reward functions, which has received great attention recently. To efficiently handle this problem, especially in the case with complicated decision sets, previous studies have proposed an efficient projection-free algorithm called Mono-Frank-Wolfe (Mono-FW) using O(T)O(T)O(T) gradient evaluations and linear optimization steps in total. However, it only attains a (1−1/e)(1-1/e)(1−1/e)-regret bound of O(T4/5)O(T^{4/5})O(T4/5). In this paper, we propose an improved projection-free algorithm, namely POBGA, which reduces the regret bound to O(T3/4)O(T^{3/4})O(T3/4) while keeping the same computational complexity as Mono-FW. Instead of modifying Mono-FW, our key idea is to make a novel combination of a projection-based algorithm called online boosting gradient ascent, an infeasible projection technique, and a blocking technique. Furthermore, we consider the decentralized setting and develop a variant of POBGA, which not only reduces the current best regret bound of efficient projection-free algorithms for this setting from O(T4/5)O(T^{4/5})O(T4/5) to O(T3/4)O(T^{3/4})O(T3/4), but also reduces the total communication complexity from O(T)O(T)O(T) to O(T)O(\sqrt{T})O(T​).

View on arXiv
Comments on this paper