While standard gradient descent methods, discussed in the previous section on advanced classical optimizers, provide a baseline for training Variational Quantum Algorithms (VQAs), they operate under an implicit assumption: that the parameter space θ is Euclidean. In this standard approach, the update rule θ(t+1)=θ(t)−η∇L(θ(t)) moves parameters in the direction of steepest descent in the parameter space. However, the actual objective of our optimization is often tied to the quantum state ∣ψ(θ)⟩ generated by the Parameterized Quantum Circuit (PQC), and the mapping from parameters θ to states ∣ψ(θ)⟩ can be highly non-linear and non-uniform. Small changes in parameters might lead to large changes in the quantum state in some regions of the parameter space, while large parameter changes might yield only small state changes elsewhere.
This observation motivates moving beyond standard gradient descent towards methods that account for the geometry of the space of quantum states itself. As introduced in Chapter 1, information geometry provides the tools to analyze the structure of statistical models, including the quantum states generated by our PQC. The Quantum Natural Gradient (QNG) leverages this geometric perspective for optimization.
The core idea behind QNG is to perform gradient descent not on the flat parameter manifold, but on the curved manifold of quantum states induced by the parameters θ. The "distance" between two infinitesimally close quantum states ∣ψ(θ)⟩ and ∣ψ(θ+dθ)⟩ is measured by the Fubini-Study metric, often referred to as the Quantum Fisher Information Matrix (QFIM) in this context.
Let ∣ψ(θ)⟩ be the state prepared by a PQC with parameters θ=(θ1,…,θM). The Fubini-Study metric tensor gij(θ) captures the infinitesimal distance squared between states resulting from parameter changes dθi and dθj. Its components are given by:
gij(θ)=Re(⟨∂iψ(θ)∣∂jψ(θ)⟩−⟨∂iψ(θ)∣ψ(θ)⟩⟨ψ(θ)∣∂jψ(θ)⟩)where ∣∂iψ(θ)⟩=∂θi∂∣ψ(θ)⟩. This metric tensor forms an M×M symmetric, positive semi-definite matrix G(θ) which quantifies how much the quantum state changes locally as we vary the parameters. It effectively tells us the sensitivity of the quantum state to parameter perturbations.
The Quantum Natural Gradient modifies the standard gradient ∇L(θ) by pre-multiplying it with the inverse of the Fubini-Study metric tensor G(θ)−1:
θ(t+1)=θ(t)−ηG(θ(t))−1∇L(θ(t))Here, η is the learning rate. This update rule performs the steepest descent step directly on the manifold of quantum states. By incorporating the geometric information via G−1, the QNG update step is invariant to the specific parameterization of the PQC. It effectively rescales the gradient components based on how much each parameter actually changes the quantum state, taking larger steps in directions where parameters have little effect on the state and smaller steps where parameters are highly sensitive.
The primary challenge in using QNG is the computation and inversion of the Fubini-Study metric tensor G(θ).
Due to the computational overhead, approximations to the full QFIM are frequently used:
Quantum software libraries like PennyLane offer functionalities for computing the QFIM and implementing QNG optimizers. However, efficiently calculating the QFIM, especially on quantum hardware, remains an active area of research.
In summary, Quantum Natural Gradient represents a sophisticated optimization technique for VQAs that incorporates the geometric structure of the quantum state space. While computationally more demanding than standard gradient descent, it offers the potential for faster convergence and more robust optimization by taking steps that are inherently adapted to the sensitivity of the quantum state to parameter changes. Its practical application often involves trade-offs between geometric accuracy and computational cost, leading to the use of various approximations.
© 2025 ApX Machine Learning