Diagnosing Training Instability: Oscillations and Divergence
Was this section helpful?
Generative Adversarial Networks, Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, Yoshua Bengio, 2014arXiv preprint arXiv:1406.2661DOI: 10.48550/arXiv.1406.2661 - Introduces the foundational GAN framework and the adversarial training objective, which is critical for understanding the source of training instabilities and the min-max game.
Improved Training of Wasserstein GANs, Ishaan Gulrajani, Faruk Ahmed, Martin Arjovsky, Vincent Dumoulin, Aaron Courville, 2017Advances in Neural Information Processing Systems (NeurIPS)DOI: 10.48550/arXiv.1704.00028 - Presents a significant improvement for GAN stability by introducing the Wasserstein distance with a gradient penalty, directly addressing common issues like vanishing/exploding gradients and mode collapse.
Deep Learning (Chapter 20: Generative Adversarial Networks), Ian Goodfellow, Yoshua Bengio, Aaron Courville, 2016 (MIT Press) - Provides a detailed academic treatment of Generative Adversarial Networks, including their theoretical foundations, training dynamics, and the inherent challenges that lead to instability.