Now that we've explored different ways to construct quantum feature maps, ϕ:x↦∣ϕ(x)⟩, a natural question arises: How effective are these maps? Not all encoding strategies are created equal. Their ability to transform classical data into quantum states with useful properties directly impacts the performance of downstream quantum machine learning algorithms. Two fundamental properties govern this effectiveness: expressibility and entangling capability. Understanding and quantifying these aspects allows us to design better feature maps tailored to specific problems and hardware.
Expressibility refers to the extent to which a quantum feature map can generate diverse quantum states across the Hilbert space as the input data x varies. Think of it as the "coverage" of the feature space. A highly expressive feature map can produce states populating a large volume or a complex manifold within the state space, while a map with low expressibility might confine all encoded data points to a small or simple region.
Why does this matter? The expressibility of the feature map fundamentally limits the complexity of the functions that a subsequent quantum model (like a quantum kernel machine or a variational classifier) can represent. If the feature map doesn't sufficiently separate or structure the data points in the Hilbert space, even the most powerful quantum algorithm built upon it might fail to find a good decision boundary or model the underlying data distribution effectively.
Quantifying expressibility is an active area of research, but several approaches exist. One common idea is to compare the distribution of states {∣ϕ(x)⟩} generated by the feature map (for a representative set of inputs x) to a uniform distribution over the Hilbert space, often approximated by the Haar measure.
Numerically, this can involve:
Another approach involves analyzing the algebraic structure generated by the gates in the feature map circuit, connecting expressibility to concepts from Lie algebra and control theory. High expressibility is often associated with the feature map circuit generating a Lie algebra that spans the entire space of operators (like su(2n) for n qubits).
However, maximum expressibility isn't always desirable. Highly expressive circuits can be difficult to train, potentially leading to issues like barren plateaus, which we will discuss in Chapter 4. There's often a trade-off between expressibility and trainability.
Entanglement, the uniquely quantum correlation between subsystems, is often cited as a potential source of quantum advantage in QML. The entangling capability of a feature map measures its tendency to generate entangled states ∣ϕ(x)⟩. A feature map consisting only of single-qubit rotations, for instance, would have zero entangling capability, as it cannot create entanglement between qubits.
The ability to generate entanglement is significant because it allows the quantum state to encode correlations between input features in ways that are classically intractable for high dimensions. This might be particularly beneficial for datasets where complex feature interactions are important.
Several measures can quantify the entanglement generated by a feature map. A practical approach often involves computing an entanglement measure for the output state ∣ϕ(x)⟩ and averaging it over a dataset {xi}. Common entanglement measures for pure states include:
The entangling capability is directly related to the presence and placement of multi-qubit gates (like CNOT or CZ) within the feature map circuit. Maps employing more entangling gates, or arranging them in specific patterns, generally exhibit higher entangling capability.
A simple 2-qubit feature map circuit. Single-qubit rotations Rx,Ry,Rz encode data x=(x0,x1) and potentially trainable parameters p0,p1. The CNOT gate (CX) is responsible for generating entanglement. The entangling capability depends on the input x and parameters p.
While distinct, expressibility and entangling capability are often intertwined. Feature maps capable of generating significant entanglement typically explore larger regions of the Hilbert space, contributing to higher expressibility. A circuit lacking entangling gates can only generate product states, severely limiting its expressibility to a small submanifold of the full Hilbert space.
However, high entangling power doesn't automatically guarantee high expressibility relevant for a specific task. It's possible to generate highly entangled states that are clustered in a way that doesn't effectively separate different classes of data. Conversely, some level of expressibility can be achieved even with limited entanglement, particularly in how single-qubit states are distributed.
The ideal balance depends on the problem. For tasks believed to require capturing complex, non-local correlations, prioritizing entangling capability might be beneficial. For others, ensuring sufficient coverage (expressibility) of the relevant part of the feature space might be more important, potentially with simpler, less entangled structures to aid trainability.
Analyzing these properties helps guide the design of feature map circuits:
Tools within quantum computing libraries like PennyLane provide functionalities to compute quantities related to expressibility (e.g., using the qml.expressibility
function, which often relies on comparing state moments to Haar random) and entanglement (e.g., qml.vn_entropy
, qml.purity
). These allow for empirical evaluation of different feature map designs.
Understanding the expressibility and entangling capability of your chosen feature map is not just a theoretical exercise. It provides insights into the potential power and limitations of your QML model, guiding architecture choices and helping to interpret results, especially when comparing quantum approaches to classical benchmarks or when diagnosing training problems. As we move towards quantum kernel methods in the next chapter, we'll see how the geometry of the feature space, shaped by these properties, directly defines the quantum kernel.
© 2025 ApX Machine Learning