82
7

αα-GAN: Convergence and Estimation Guarantees

Abstract

We prove a two-way correspondence between the min-max optimization of general CPE loss function GANs and the minimization of associated ff-divergences. We then focus on α\alpha-GAN, defined via the α\alpha-loss, which interpolates several GANs (Hellinger, vanilla, Total Variation) and corresponds to the minimization of the Arimoto divergence. We show that the Arimoto divergences induced by α\alpha-GAN equivalently converge, for all αR>0{}\alpha\in \mathbb{R}_{>0}\cup\{\infty\}. However, under restricted learning models and finite samples, we provide estimation bounds which indicate diverse GAN behavior as a function of α\alpha. Finally, we present empirical results on a toy dataset that highlight the practical utility of tuning the α\alpha hyperparameter.

View on arXiv
Comments on this paper