While the core idea of Generative Adversarial Networks (GANs) involves a generator and a discriminator competing in a minimax game, achieving stable and effective training in practice often requires specific techniques. Standard GAN training can suffer from issues like mode collapse, where the generator produces only a limited variety of outputs, or vanishing/exploding gradients, making convergence difficult.
This chapter focuses on understanding these training difficulties and introduces methods designed to improve stability. We will examine:
You will learn the motivation behind these techniques and how they modify the training process to foster better convergence and prevent common failure modes. We will also cover the practical implementation details, preparing you to apply these methods in your own GAN projects.
3.1 The Challenge of Non-Convergence
3.2 Mode Collapse: Causes and Consequences
3.3 Alternative Divergences: Wasserstein Distance
3.4 Weight Clipping in WGAN
3.5 Gradient Penalty (WGAN-GP)
3.6 Spectral Normalization
3.7 Two Time-Scale Update Rule (TTUR)
3.8 Relativistic GANs
3.9 Implementing WGAN-GP: Practice
© 2025 ApX Machine Learning