Classical generative models often construct an explicit probability distribution or use intricate architectures like GANs or VAEs. Quantum Circuit Born Machines (QCBMs) offer a different approach, harnessing the inherent probabilistic nature of quantum mechanics. The core idea is surprisingly direct: a parameterized quantum circuit (PQC) can be used to prepare a quantum state whose measurement statistics define the probability distribution we want to model.
Recall the Born rule from fundamental quantum mechanics: if a quantum system is in state ∣ψ⟩, the probability of measuring a specific outcome x (represented by the basis state ∣x⟩) is given by P(x)=∣⟨x∣ψ⟩∣2.
A QCBM leverages this principle directly. We start with an easily prepared initial state, typically the all-zeros state ∣0⟩⊗n for n qubits. We then apply a PQC, denoted by a unitary operation U(θ), where θ represents the set of trainable circuit parameters (like rotation angles). This transforms the initial state into a parameterized state: ∣ψ(θ)⟩=U(θ)∣0⟩⊗n
Measuring this state ∣ψ(θ)⟩ in the computational basis (the basis corresponding to classical bitstrings like ∣001⟩, ∣101⟩, etc.) yields an outcome x (a specific bitstring) with probability: pθ(x)=∣⟨x∣ψ(θ)⟩∣2=∣⟨x∣U(θ)∣0⟩⊗n∣2
This equation defines the probability distribution pθ(x) implicitly modeled by the QCBM. The circuit U(θ) acts as the generator, and the distribution arises naturally from quantum measurement. The goal of training is to adjust the parameters θ such that pθ(x) closely approximates the target data distribution pdata(x).
Basic workflow of a Quantum Circuit Born Machine (QCBM). Trainable parameters θ control a Parameterized Quantum Circuit (PQC) U(θ) applied to an initial state ∣0⟩⊗n. Measuring the resulting state ∣ψ(θ)⟩ yields samples x according to the probability distribution pθ(x).
Since the QCBM defines a probability distribution pθ(x), we can train it similarly to classical probabilistic models. We need a way to quantify the difference between the distribution generated by the QCBM, pθ(x), and the target distribution derived from the training data, pdata(x). Common approaches include minimizing a divergence or distance metric:
Kullback-Leibler (KL) Divergence: Measures the difference between two probability distributions. The objective is to minimize KL(pdata∣∣pθ). Calculating this often requires estimating both distributions. For discrete data, pdata can be estimated from frequencies in the dataset. Evaluating pθ(x) for all possible x can be computationally intensive, especially for many qubits, as it requires calculating ∣⟨x∣U(θ)∣0⟩⊗n∣2 for potentially 2n outcomes. Approximations or alternative cost functions are often necessary.
Maximum Mean Discrepancy (MMD): MMD measures the distance between distributions based on the difference in the mean embeddings of samples in a Reproducing Kernel Hilbert Space (RKHS). A common choice is the Gaussian kernel. The MMD loss compares statistics of samples drawn from pdata (the real data) and samples generated by the QCBM (pθ). An advantage is that MMD can often be estimated efficiently using batches of samples from both distributions, avoiding the need to explicitly compute pθ(x) for all x.
Maximum Likelihood Estimation (MLE): If we can efficiently evaluate pθ(x) for specific data points xi from the training set, we can aim to maximize the log-likelihood ∑ilogpθ(xi). However, calculating individual probabilities pθ(xi) can still be demanding.
Regardless of the chosen cost function C(θ), training involves finding the parameters θ∗ that minimize it: θ∗=argminθC(θ) This optimization is typically performed using gradient-based methods. Gradients ∇θC(θ) can be estimated using techniques like the parameter-shift rule (discussed in Chapter 4 for VQAs) or finite differences, often requiring multiple circuit executions per gradient component per optimization step. Classical optimizers like Adam or SGD are then used to update the parameters θ.
The choice of the PQC architecture U(θ) is significant for the QCBM's performance.
Once the QCBM is trained (i.e., we have found suitable parameters θ∗), generating new samples that resemble the training data is straightforward:
QCBMs represent a fundamental quantum approach to generative modeling. They directly map the structure of a quantum circuit and the physics of measurement to the task of learning and sampling from probability distributions. While challenges in calculating cost functions, estimating gradients, and ensuring trainability exist, they offer a potentially powerful alternative to classical methods, especially for distributions that might have complex correlations naturally captured by quantum states.
© 2025 ApX Machine Learning