228

High-Dimensional Calibration from Swap Regret

Main:9 Pages
4 Figures
Bibliography:3 Pages
Appendix:14 Pages
Abstract

We study the online calibration of multi-dimensional forecasts over an arbitrary convex set PRd\mathcal{P} \subset \mathbb{R}^d relative to an arbitrary norm \Vert\cdot\Vert. We connect this with the problem of external regret minimization for online linear optimization, showing that if it is possible to guarantee O(ρT)O(\sqrt{\rho T}) worst-case regret after TT rounds when actions are drawn from P\mathcal{P} and losses are drawn from the dual \Vert \cdot \Vert_* unit norm ball, then it is also possible to obtain ϵ\epsilon-calibrated forecasts after T=exp(O(ρ/ϵ2))T = \exp(O(\rho /\epsilon^2)) rounds. When P\mathcal{P} is the dd-dimensional simplex and \Vert \cdot \Vert is the 1\ell_1-norm, the existence of O(Tlogd)O(\sqrt{T\log d})-regret algorithms for learning with experts implies that it is possible to obtain ϵ\epsilon-calibrated forecasts after T=exp(O(logd/ϵ2))=dO(1/ϵ2)T = \exp(O(\log{d}/\epsilon^2)) = d^{O(1/\epsilon^2)} rounds, recovering a recent result of Peng (2025).Interestingly, our algorithm obtains this guarantee without requiring access to any online linear optimization subroutine or knowledge of the optimal rate ρ\rho -- in fact, our algorithm is identical for every setting of P\mathcal{P} and \Vert \cdot \Vert. Instead, we show that the optimal regularizer for the above OLO problem can be used to upper bound the above calibration error by a swap regret, which we then minimize by running the recent TreeSwap algorithm with Follow-The-Leader as a subroutine.Finally, we prove that any online calibration algorithm that guarantees ϵT\epsilon T 1\ell_1-calibration error over the dd-dimensional simplex requires Texp(poly(1/ϵ))T \geq \exp(\mathrm{poly}(1/\epsilon)) (assuming dpoly(1/ϵ)d \geq \mathrm{poly}(1/\epsilon)). This strengthens the corresponding dΩ(log1/ϵ)d^{\Omega(\log{1/\epsilon})} lower bound of Peng, and shows that an exponential dependence on 1/ϵ1/\epsilon is necessary.

View on arXiv
Comments on this paper