95
v1v2 (latest)

Phase transition of the Sinkhorn-Knopp algorithm

2 Figures
Appendix:44 Pages
Abstract

The matrix scaling problem, particularly the Sinkhorn-Knopp algorithm, has been studied for over 60 years. In practice, the algorithm often yields high-quality approximations within just a few iterations. Theoretically, however, the best-known upper bound places it in the class of pseudopolynomial-time approximation algorithms. Meanwhile, the lower-bound landscape remains largely unexplored. Two fundamental questions persist: what accounts for the algorithm's strong empirical performance, and can a tight bound on its iteration count be established?For an n×nn\times n matrix, its normalized version is obtained by dividing each entry by its largest entry. We say that a normalized matrix has a density γ\gamma if there exists a constant ρ>0\rho > 0 such that one row or column has exactly γn\lceil \gamma n \rceil entries with values at least ρ\rho, and every other row and column has at least γn\lceil \gamma n \rceil such entries.For the upper bound, we show that the Sinkhorn-Knopp algorithm produces a nearly doubly stochastic matrix in O(lognlogε)O(\log n - \log \varepsilon) iterations and O~(n2)\widetilde{O}(n^2) time for all nonnegative square matrices whose normalized version has a density γ>1/2\gamma > 1/2. Such matrices cover both the algorithm's principal practical inputs and its typical theoretical regime, and the O~(n2)\widetilde{O}(n^2) runtime is optimal.For the lower bound, we establish a tight bound of Ω~(n1/2/ε)\widetilde{\Omega}\left(n^{1/2}/\varepsilon\right) iterations for positive matrices under the 2\ell_2-norm error measure. Moreover, for every γ<1/2\gamma < 1/2, there exists a matrix with density γ\gamma for which the algorithm requires Ω(n1/2/ε)\Omega\left(n^{1/2}/\varepsilon\right) iterations.In summary, our results reveal a sharp phase transition in the Sinkhorn-Knopp algorithm at the density threshold γ=1/2\gamma = 1/2.

View on arXiv
Comments on this paper