Effective Dynamics of Generative Adversarial Networks

ORAL

Abstract

Generative adversarial networks (GANs) are a family of machine-learning models that use adversarial training to generate new samples with the same statistics as the training samples. Mode collapse is a major form of training failure in which the generator fails to reproduce the full diversity of modes in the training data. Here, we present a simplified model of GAN training dynamics, which allows us to study the conditions under which mode collapse occurs. Our effective model replaces the generator neural network with a collection of particles in the output space; particles interact with the target distribution only via the discriminator and mutually couple due to the generator. Our model GAN reveals a mode-collapse transition, the shape of which can be related to the type of discriminator through the frequency principle. Further, we find that gradient regularizers of intermediate strengths can optimally yield convergence through critical damping of the generator dynamics. Our effective GAN model thus provides an interpretable physical framework for understanding and improving adversarial training.

* S. W. is grateful for support from an NSF CAREER Award (Grant No. PHY-2146581).

Publication: S. Durr, Y. Mroueh, Y. Tu, and S. Wang. Effective dynamics of generative adversarial networks. Physical Review X 13, 041004 (2023).

Presenters

  • Shenshen Wang

    University of California, Los Angeles

Authors

  • Shenshen Wang

    University of California, Los Angeles

  • Steven Durr

    University of California, Los Angeles

  • Youssef Mroueh

    IBM T. J. Watson Research Center

  • Yuhai Tu

    IBM T. J. Watson Research Center