ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1901.02975
16
9

A witness function based construction of discriminative models using Hermite polynomials

10 January 2019
H. Mhaskar
A. Cloninger
Xiuyuan Cheng
ArXivPDFHTML
Abstract

In machine learning, we are given a dataset of the form {(xj,yj)}j=1M\{(\mathbf{x}_j,y_j)\}_{j=1}^M{(xj​,yj​)}j=1M​, drawn as i.i.d. samples from an unknown probability distribution μ\muμ; the marginal distribution for the xj\mathbf{x}_jxj​'s being μ∗\mu^*μ∗. We propose that rather than using a positive kernel such as the Gaussian for estimation of these measures, using a non-positive kernel that preserves a large number of moments of these measures yields an optimal approximation. We use multi-variate Hermite polynomials for this purpose, and prove optimal and local approximation results in a supremum norm in a probabilistic sense. Together with a permutation test developed with the same kernel, we prove that the kernel estimator serves as a `witness function' in classification problems. Thus, if the value of this estimator at a point x\mathbf{x}x exceeds a certain threshold, then the point is reliably in a certain class. This approach can be used to modify pretrained algorithms, such as neural networks or nonlinear dimension reduction techniques, to identify in-class vs out-of-class regions for the purposes of generative models, classification uncertainty, or finding robust centroids. This fact is demonstrated in a number of real world data sets including MNIST, CIFAR10, Science News documents, and LaLonde data sets.

View on arXiv
Comments on this paper