195
v1v2v3 (latest)

Near-Optimal Algorithms for Omniprediction

Main:62 Pages
Bibliography:5 Pages
1 Tables
Appendix:16 Pages
Abstract

Omnipredictors are simple prediction functions that encode loss-minimizing predictions with respect to a hypothesis class HH, simultaneously for every loss function within a class of losses LL. In this work, we give near-optimal learning algorithms for omniprediction, in both the online and offline settings. To begin, we give an oracle-efficient online learning algorithm that acheives (L,H)(L,H)-omniprediction with O~(TlogH)\tilde O (\sqrt{T \log |H|}) regret for any class of Lipschitz loss functions LLLipL \subseteq L_\mathrm{Lip}. Quite surprisingly, this regret bound matches the optimal regret for \emph{minimization of a single loss function} (up to a log(T)\sqrt{\log(T)} factor). Given this online algorithm, we develop an online-to-offline conversion that achieves near-optimal complexity across a number of measures. In particular, for all bounded loss functions within the class of Bounded Variation losses LBVL_\mathrm{BV} (which include all convex, all Lipschitz, and all proper losses) and any (possibly-infinite) HH, we obtain an offline learning algorithm that, leveraging an (offline) ERM oracle and mm samples from DD, returns an efficient (LBV,H,ϵ(m))(L_{\mathrm{BV}},H,\epsilon(m))-omnipredictor for ε(m)\varepsilon(m) scaling near-linearly in the Rademacher complexity of a class derived from HH by taking convex combinations of a fixed number of elements of ThH\mathrm{Th} \circ H.

View on arXiv
Comments on this paper