ResearchTrend.AI
  • Papers
  • Communities
  • Organizations
  • Events
  • Blog
  • Pricing
  • Feedback
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1905.04039
29
1

Optimal rates for F-score binary classification

10 May 2019
Evgenii Chzhen
ArXiv (abs)PDFHTML
Abstract

We study the minimax settings of binary classification with F-score under the β\betaβ-smoothness assumptions on the regression function η(x)=P(Y=1∣X=x)\eta(x) = \mathbb{P}(Y = 1|X = x)η(x)=P(Y=1∣X=x) for x∈Rdx \in \mathbb{R}^dx∈Rd. We propose a classification procedure which under the α\alphaα-margin assumption achieves the rate O(n−−(1+α)β/(2β+d))O(n^{--(1+\alpha)\beta/(2\beta+d)})O(n−−(1+α)β/(2β+d)) for the excess F-score. In this context, the Bayes optimal classifier for the F-score can be obtained by thresholding the aforementioned regression function η\etaη on some level θ∗\theta^*θ∗ to be estimated. The proposed procedure is performed in a semi-supervised manner, that is, for the estimation of the regression function we use a labeled dataset of size n∈Nn \in \mathbb{N}n∈N and for the estimation of the optimal threshold θ∗\theta^*θ∗ we use an unlabeled dataset of size N∈NN \in \mathbb{N}N∈N. Interestingly, the value of N∈NN \in \mathbb{N}N∈N does not affect the rate of convergence, which indicates that it is "harder" to estimate the regression function η\etaη than the optimal threshold θ∗\theta^*θ∗. This further implies that the binary classification with F-score behaves similarly to the standard settings of binary classification. Finally, we show that the rates achieved by the proposed procedure are optimal in the minimax sense up to a constant factor.

View on arXiv
Comments on this paper