154
v1v2 (latest)

On the Universal Near Optimality of Hedge in Combinatorial Settings

Main:9 Pages
1 Figures
Bibliography:3 Pages
Appendix:16 Pages
Abstract

In this paper, we study the classical Hedge algorithm in combinatorial settings. In each round, the learner selects a vector xt\boldsymbol{x}_t from a set X{0,1}dX \subseteq \{0,1\}^d, observes a full loss vector ytRd\boldsymbol{y}_t \in \mathbb{R}^d, and incurs a loss xt,yt[1,1]\langle \boldsymbol{x}_t, \boldsymbol{y}_t \rangle \in [-1,1]. This setting captures several important problems, including extensive-form games, resource allocation, mm-sets, online multitask learning, and shortest-path problems on directed acyclic graphs (DAGs). It is well known that Hedge achieves a regret of O(TlogX)O\big(\sqrt{T \log |X|}\big) after TT rounds of interaction. In this paper, we ask whether Hedge is optimal across all combinatorial settings. To that end, we show that for any X{0,1}dX \subseteq \{0,1\}^d, Hedge is near-optimal--specifically, up to a logd\sqrt{\log d} factor--by establishing a lower bound of Ω(Tlog(X)/logd)\Omega\big(\sqrt{T \log(|X|)/\log d}\big) that holds for any algorithm. We then identify a natural class of combinatorial sets--namely, mm-sets with logdmd\log d \leq m \leq \sqrt{d}--for which this lower bound is tight, and for which Hedge is provably suboptimal by a factor of exactly logd\sqrt{\log d}. At the same time, we show that Hedge is optimal for online multitask learning, a generalization of the classical KK-experts problem. Finally, we leverage the near-optimality of Hedge to establish the existence of a near-optimal regularizer for online shortest-path problems in DAGs--a setting that subsumes a broad range of combinatorial domains. Specifically, we show that the classical Online Mirror Descent (OMD) algorithm, when instantiated with the dilated entropy regularizer, is iterate-equivalent to Hedge, and therefore inherits its near-optimal regret guarantees for DAGs.

View on arXiv
Comments on this paper