Timezone: »

Data-Efficient GAN Training Beyond (Just) Augmentations: A Lottery Ticket Perspective
Tianlong Chen · Yu Cheng · Zhe Gan · Jingjing Liu · Zhangyang Wang

Wed Dec 08 04:30 PM -- 06:00 PM (PST) @

Training generative adversarial networks (GANs) with limited real image data generally results in deteriorated performance and collapsed models. To conquer this challenge, we are inspired by the latest observation, that one can discover independently trainable and highly sparse subnetworks (a.k.a., lottery tickets) from GANs. Treating this as an inductive prior, we suggest a brand-new angle towards data-efficient GAN training: by first identifying the lottery ticket from the original GAN using the small training set of real images; and then focusing on training that sparse subnetwork by re-using the same set. We find our coordinated framework to offer orthogonal gains to existing real image data augmentation methods, and we additionally present a new feature-level augmentation that can be applied together with them. Comprehensive experiments endorse the effectiveness of our proposed framework, across various GAN architectures (SNGAN, BigGAN, and StyleGAN-V2) and diverse datasets (CIFAR-10, CIFAR-100, Tiny-ImageNet, ImageNet, and multiple few-shot generation datasets). Codes are available at: https://github.com/VITA-Group/Ultra-Data-Efficient-GAN-Training.

Author Information

Tianlong Chen (Unversity of Texas at Austin)
Yu Cheng (Microsoft Research)
Zhe Gan (Duke University)
Jingjing Liu (Microsoft)
Zhangyang Wang (UT Austin)

More from the Same Authors