ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1508.04905
18
2
v1v2 (latest)

Theoretical analysis of cross-validation for estimating the risk of the k-Nearest Neighbor classifier

20 August 2015
Alain Celisse
T. Mary-Huard
ArXiv (abs)PDFHTML
Abstract

The present work aims at deriving theoretical guaranties on the behavior of some cross-validation procedures applied to the kkk-nearest neighbors (kkkNN) rule in the context of binary classification. Here we focus on the leave-ppp-out cross-validation (LpppO) used to assess the performance of the kkkNN classifier. Remarkably this LpppO estimator can be efficiently computed in this context using closed-form formulas derived by \cite{CelisseMaryHuard11}. We describe a general strategy to derive moment and exponential concentration inequalities for the LpppO estimator applied to the kkkNN classifier. Such results are obtained first by exploiting the connection between the LpppO estimator and U-statistics, and second by making an intensive use of the generalized Efron-Stein inequality applied to the L111O estimator. One other important contribution is made by deriving new quantifications of the discrepancy between the LpppO estimator and the classification error/risk of the kkkNN classifier. The optimality of these bounds is discussed by means of several lower bounds as well as simulation experiments.

View on arXiv
Comments on this paper