ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1912.00953
8
88

LOGAN: Latent Optimisation for Generative Adversarial Networks

2 December 2019
Y. Wu
Jeff Donahue
David Balduzzi
Karen Simonyan
Timothy Lillicrap
    GAN
ArXivPDFHTML
Abstract

Training generative adversarial networks requires balancing of delicate adversarial dynamics. Even with careful tuning, training may diverge or end up in a bad equilibrium with dropped modes. In this work, we improve CS-GAN with natural gradient-based latent optimisation and show that it improves adversarial dynamics by enhancing interactions between the discriminator and the generator. Our experiments demonstrate that latent optimisation can significantly improve GAN training, obtaining state-of-the-art performance for the ImageNet (128×128128 \times 128128×128) dataset. Our model achieves an Inception Score (IS) of 148148148 and an Fr\échet Inception Distance (FID) of 3.43.43.4, an improvement of 17%17\%17% and 32%32\%32% in IS and FID respectively, compared with the baseline BigGAN-deep model with the same architecture and number of parameters.

View on arXiv
Comments on this paper