Before we examine quantum approaches to generative modeling, let's briefly review two influential classical frameworks: Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs). Understanding their core ideas provides valuable context for their quantum counterparts. The goal in both cases is to learn a model distribution pmodel(x) that approximates an unknown data distribution pdata(x), allowing us to generate new samples resembling the original data.
GANs, introduced by Ian Goodfellow and colleagues in 2014, employ a clever adversarial training process involving two neural networks:
The training process is formulated as a minimax game. The Discriminator is trained to maximize its accuracy in distinguishing real samples from fake samples generated by G. Simultaneously, the Generator is trained to minimize the probability that the Discriminator correctly identifies its outputs as fake. Essentially, G tries to fool D.
The objective function for this game is typically expressed as:
GminDmaxV(D,G)=Ex∼pdata(x)[logD(x)]+Ez∼pz(z)[log(1−D(G(z)))]Here, E denotes expectation. The Discriminator D wants to maximize this value (making D(x) close to 1 for real x and D(G(z)) close to 0 for fake samples). The Generator G wants to minimize this value (by making D(G(z)) close to 1, effectively fooling the Discriminator).
Basic architecture of a Generative Adversarial Network (GAN), illustrating the interaction between the Generator and Discriminator.
Training GANs can be challenging due to potential instabilities like mode collapse (where the generator produces only a limited variety of samples) or vanishing gradients. However, when successful, GANs are known for generating high-fidelity samples, particularly in domains like image generation.
VAEs, introduced by Diederik Kingma and Max Welling in 2013, approach generative modeling from a probabilistic perspective, combining ideas from variational inference and autoencoders. A VAE consists of two main parts, often implemented as neural networks:
The goal of training a VAE is to maximize the log-likelihood of the data, logp(x). Since this is often intractable, VAEs maximize a lower bound instead, known as the Evidence Lower Bound (ELBO):
L(θ,ϕ;x)=Ez∼qϕ(z∣x)[logpθ(x∣z)]−DKL(qϕ(z∣x)∣∣p(z))The ELBO consists of two terms:
A specific technique called the "reparameterization trick" is used to allow gradients to flow back through the sampling process within the encoder, enabling training via standard backpropagation.
Basic architecture of a Variational Autoencoder (VAE), showing the encoding, latent sampling, and decoding stages.
VAEs typically provide more stable training compared to GANs and offer an explicit probabilistic model with a well-defined likelihood objective (the ELBO). However, the samples generated by VAEs are sometimes perceived as less sharp or realistic than those from state-of-the-art GANs, potentially due to the simplifying assumptions made (e.g., Gaussian posteriors) and the influence of the reconstruction term.
These classical models establish fundamental concepts in generative modeling: representing distributions implicitly (GANs) or explicitly via latent variables (VAEs), defining objective functions, and navigating training dynamics. As we move into the quantum domain, we'll see how quantum circuits can be used as components within similar frameworks, potentially offering new ways to capture complex data distributions.
© 2025 ApX Machine Learning