42
2

Revisiting Projection-Free Online Learning with Time-Varying Constraints

Abstract

We investigate constrained online convex optimization, in which decisions must belong to a fixed and typically complicated domain, and are required to approximately satisfy additional time-varying constraints over the long term. In this setting, the commonly used projection operations are often computationally expensive or even intractable. To avoid the time-consuming operation, several projection-free methods have been proposed with an O(T3/4logT)\mathcal{O}(T^{3/4} \sqrt{\log T}) regret bound and an O(T7/8)\mathcal{O}(T^{7/8}) cumulative constraint violation (CCV) bound for general convex losses. In this paper, we improve this result and further establish \textit{novel} regret and CCV bounds when loss functions are strongly convex. The primary idea is to first construct a composite surrogate loss, involving the original loss and constraint functions, by utilizing the Lyapunov-based technique. Then, we propose a parameter-free variant of the classical projection-free method, namely online Frank-Wolfe (OFW), and run this new extension over the online-generated surrogate loss. Theoretically, for general convex losses, we achieve an O(T3/4)\mathcal{O}(T^{3/4}) regret bound and an O(T3/4logT)\mathcal{O}(T^{3/4} \log T) CCV bound, both of which are order-wise tighter than existing results. For strongly convex losses, we establish new guarantees of an O(T2/3)\mathcal{O}(T^{2/3}) regret bound and an O(T5/6)\mathcal{O}(T^{5/6}) CCV bound. Moreover, we also extend our methods to a more challenging setting with bandit feedback, obtaining similar theoretical findings. Empirically, experiments on real-world datasets have demonstrated the effectiveness of our methods.

View on arXiv
Comments on this paper