ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2112.13867
15
0

Depth and Feature Learning are Provably Beneficial for Neural Network Discriminators

27 December 2021
Carles Domingo-Enrich
    MLT
    MDE
ArXivPDFHTML
Abstract

We construct pairs of distributions μd,νd\mu_d, \nu_dμd​,νd​ on Rd\mathbb{R}^dRd such that the quantity ∣Ex∼μd[F(x)]−Ex∼νd[F(x)]∣|\mathbb{E}_{x \sim \mu_d} [F(x)] - \mathbb{E}_{x \sim \nu_d} [F(x)]|∣Ex∼μd​​[F(x)]−Ex∼νd​​[F(x)]∣ decreases as Ω(1/d2)\Omega(1/d^2)Ω(1/d2) for some three-layer ReLU network FFF with polynomial width and weights, while declining exponentially in ddd if FFF is any two-layer network with polynomial weights. This shows that deep GAN discriminators are able to distinguish distributions that shallow discriminators cannot. Analogously, we build pairs of distributions μd,νd\mu_d, \nu_dμd​,νd​ on Rd\mathbb{R}^dRd such that ∣Ex∼μd[F(x)]−Ex∼νd[F(x)]∣|\mathbb{E}_{x \sim \mu_d} [F(x)] - \mathbb{E}_{x \sim \nu_d} [F(x)]|∣Ex∼μd​​[F(x)]−Ex∼νd​​[F(x)]∣ decreases as Ω(1/(dlog⁡d))\Omega(1/(d\log d))Ω(1/(dlogd)) for two-layer ReLU networks with polynomial weights, while declining exponentially for bounded-norm functions in the associated RKHS. This confirms that feature learning is beneficial for discriminators. Our bounds are based on Fourier transforms.

View on arXiv
Comments on this paper