359
v1v2 (latest)

Universal consistency and rates of convergence of multiclass prototype algorithms in metric spaces

Journal of machine learning research (JMLR), 2020
Abstract

We study universal consistency and convergence rates of simple nearest-neighbor prototype rules for the problem of multiclass classification in metric paces. We first show that a novel data-dependent partitioning rule, named Proto-NN, is universally consistent in any metric space that admits a universally consistent rule. Proto-NN is a significant simplification of OptiNet, a recently proposed compression-based algorithm that, to date, was the only algorithm known to be universally consistent in such a general setting. Practically, Proto-NN is simpler to implement and enjoys reduced computational complexity. We then proceed to study convergence rates of the excess error probability. We first obtain rates for the standard kk-NN rule under a margin condition and a new generalized-Lipschitz condition. The latter is an extension of a recently proposed modified-Lipschitz condition from Rd\mathbb R^d to metric spaces. Similarly to the modified-Lipschitz condition, the new condition avoids any boundness assumptions on the data distribution. While obtaining rates for Proto-NN is left open, we show that a second prototype rule that hybridizes between kk-NN and Proto-NN achieves the same rates as kk-NN while enjoying similar computational advantages as Proto-NN. However, as kk-NN, this hybrid rule is not consistent in general.

View on arXiv
Comments on this paper