Gang of GANs: Generative Adversarial Networks with Maximum Margin Ranking

People

Felix Juefei Xu (Contact)

Vishnu Naresh Boddeti (Contact)

Marios Savvides

Carnegie Mellon University


Code

GoGAN (PyTorch), to be released.


Abstract

Traditional generative adversarial networks (GAN) and many of its variants are trained by minimizing the KL or JS-divergence loss that measures how close the generated data distribution is from the true data distribution. A recent advance called the WGAN based on Wasserstein distance can improve on the KL and JS-divergence based GANs, and alleviate the gradient vanishing, instability, and mode collapse issues that are common in the GAN training. In this work, we aim at improving on the WGAN by first generalizing its discriminator loss to a margin-based one, which leads to a better discriminator, and in turn a better generator, and then carrying out a progressive training paradigm involving multiple GANs to contribute to the maximum margin ranking loss so that the GAN at later stages will improve upon early stages. We call this method Gang of GANs (GoGAN). We have shown theoretically that the proposed GoGAN can reduce the gap between the true data distribution and the generated data distribution by at least half in an optimally trained WGAN. We have also proposed a new way of measuring GAN quality which is based on image completion tasks. We have evaluated our method on three visual datasets: CelebA, LSUN Bedroom, and CIFAR-10, and have seen both visual and quantitative improvement over baseline WGAN.


Overview

Our GoGAN allow GANs at later stages to improve upon early stages in a progressive training paradigm.

Discriminator scores ordering.


Contributions


References