20
2

Wasserstein GANs are Minimax Optimal Distribution Estimators

Abstract

We provide non asymptotic rates of convergence of the Wasserstein Generative Adversarial networks (WGAN) estimator. We build neural networks classes representing the generators and discriminators which yield a GAN that achieves the minimax optimal rate for estimating a certain probability measure μ\mu with support in Rp\mathbb{R}^p. The probability μ\mu is considered to be the push forward of the Lebesgue measure on the dd-dimensional torus Td\mathbb{T}^d by a map g:TdRpg^\star:\mathbb{T}^d\rightarrow \mathbb{R}^p of smoothness β+1\beta+1. Measuring the error with the γ\gamma-H\"older Integral Probability Metric (IPM), we obtain up to logarithmic factors, the minimax optimal rate O(nβ+γ2β+dn12)O(n^{-\frac{\beta+\gamma}{2\beta +d}}\vee n^{-\frac{1}{2}}) where nn is the sample size, β\beta determines the smoothness of the target measure μ\mu, γ\gamma is the smoothness of the IPM (γ=1\gamma=1 is the Wasserstein case) and dpd\leq p is the intrinsic dimension of μ\mu. In the process, we derive a sharp interpolation inequality between H\"older IPMs. This novel result of theory of functions spaces generalizes classical interpolation inequalities to the case where the measures involved have densities on different manifolds.

View on arXiv
Comments on this paper