Generative Adversarial Nets, Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, Yoshua Bengio, 2014Advances in Neural Information Processing Systems, Vol. 27 (Curran Associates, Inc.)DOI: 10.48550/arXiv.1406.2661 - Introduces the original GAN framework, defining the minimax objective and its connection to Jensen-Shannon divergence, which are foundational to understanding mode collapse.
Wasserstein GAN, Martin Arjovsky, Soumith Chintala, and Léon Bottou, 2017Proceedings of the 34th International Conference on Machine Learning, Vol. 70 (PMLR) - Proposes using the Wasserstein-1 distance as a GAN loss function to provide more stable gradients and address mode collapse and vanishing gradient issues.
Improved Training of Wasserstein GANs, Ishaan Gulrajani, Faruk Ahmed, Martin Arjovsky, Vincent Dumoulin, Aaron C. Courville, 2017Advances in Neural Information Processing Systems 30 - Introduces a gradient penalty to enforce the Lipschitz constraint in WGANs, significantly improving training stability and sample quality.
Improved Techniques for Training GANs, Tim Salimans, Ian Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, Xi Chen, 2016Advances in Neural Information Processing Systems, Vol. 29 (NeurIPS)DOI: 10.48550/arXiv.1606.03498 - Presents several techniques, including minibatch discrimination, to mitigate mode collapse and stabilize GAN training.