Standard Generative Adversarial Networks produce samples reflecting the training data distribution, but offer limited direct control over the specific output. This chapter introduces techniques for guiding the generation process. You will learn to implement Conditional GANs (cGANs), which generate data matching specific attributes or labels, often denoted as y. We cover how to integrate this conditional information y into both the generator G(z,y) and discriminator D(x,y).
We then study Information Maximizing GANs (InfoGAN), a method for learning interpretable latent codes c in an unsupervised fashion by maximizing mutual information I(c;G(z,c)) between these codes and the generator output. The chapter also addresses methods for manipulating the learned latent space z to modify output characteristics post-generation and discusses approaches like StackGAN for tasks such as text-to-image synthesis. Finally, we consider the concept and measurement of disentanglement in learned representations. Hands-on sections provide practice in building these conditional models.
4.1 Introduction to Conditional GANs (cGANs)
4.2 Architectures for cGANs
4.3 Information Maximizing GANs (InfoGAN)
4.4 StackGAN: Text-to-Image Synthesis
4.5 Controlling Attributes via Latent Space Manipulation
4.6 Disentanglement Metrics and Challenges
4.7 Building a Conditional GAN: Hands-on Practical
© 2025 ApX Machine Learning