ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2205.11470
28
8

Exploiting the Curvature of Feasible Sets for Faster Projection-Free Online Learning

23 May 2022
Zakaria Mhammedi
ArXivPDFHTML
Abstract

In this paper, we develop new efficient projection-free algorithms for Online Convex Optimization (OCO). Online Gradient Descent (OGD) is an example of a classical OCO algorithm that guarantees the optimal O(T)O(\sqrt{T})O(T​) regret bound. However, OGD and other projection-based OCO algorithms need to perform a Euclidean projection onto the feasible set C⊂Rd\mathcal{C}\subset \mathbb{R}^dC⊂Rd whenever their iterates step outside C\mathcal{C}C. For various sets of interests, this projection step can be computationally costly, especially when the ambient dimension is large. This has motivated the development of projection-free OCO algorithms that swap Euclidean projections for often much cheaper operations such as Linear Optimization (LO). However, state-of-the-art LO-based algorithms only achieve a suboptimal O(T3/4)O(T^{3/4})O(T3/4) regret for general OCO. In this paper, we leverage recent results in parameter-free Online Learning, and develop an OCO algorithm that makes two calls to an LO Oracle per round and achieves the near-optimal O~(T)\widetilde{O}(\sqrt{T})O(T​) regret whenever the feasible set is strongly convex. We also present an algorithm for general convex sets that makes O~(d)\widetilde O(d)O(d) expected number of calls to an LO Oracle per round and guarantees a O~(T2/3)\widetilde O(T^{2/3})O(T2/3) regret, improving on the previous best O(T3/4)O(T^{3/4})O(T3/4). We achieve the latter by approximating any convex set C\mathcal{C}C by a strongly convex one, where LO can be performed using O~(d)\widetilde {O}(d)O(d) expected number of calls to an LO Oracle for C\mathcal{C}C.

View on arXiv
Comments on this paper