Support Vector Machines (SVMs) are powerful supervised learning models used for classification and regression tasks. A primary strength of SVMs lies in their use of the "kernel trick." Instead of explicitly mapping data into a high-dimensional feature space, SVMs rely on a kernel function k(x,x′) that computes the inner product of data points x and x′ in that feature space. This allows SVMs to find complex, non-linear decision boundaries efficiently.
Classical SVMs utilize kernels like the linear, polynomial, or Radial Basis Function (RBF) kernel. The core idea behind the Quantum Support Vector Machine (QSVM) is straightforward: replace the classical kernel function with a quantum kernel computed using a quantum feature map. As discussed earlier in this chapter, a quantum feature map ϕ encodes classical data x into a quantum state ∣ϕ(x)⟩ residing in a potentially vast Hilbert space H. The quantum kernel is then defined based on the similarity between these quantum states, typically as the squared overlap (fidelity):
kq(x,x′)=∣⟨ϕ(x)∣ϕ(x′)⟩∣2Other definitions based on the inner product ⟨ϕ(x)∣ϕ(x′)⟩ are also sometimes used, ensuring the resulting kernel matrix is positive semi-definite, a requirement for the underlying SVM optimization problem.
The motivation for QSVM stems from the hypothesis that quantum feature maps can map classical data into Hilbert spaces where the data becomes more easily separable, potentially offering an advantage over classical kernels for certain datasets. The structure of the quantum feature space H, influenced by entanglement and quantum interference generated by the feature map circuit Uϕ(x), might capture complex correlations in the data that are inaccessible to standard classical kernels. If a quantum feature map provides a representation where a linear separator (in H) performs significantly better than what's achievable with classical kernels in their respective feature spaces, QSVM could outperform classical SVMs.
However, proving such an advantage is non-trivial and depends heavily on the choice of the feature map and the specific problem domain. The potential for quantum advantage here lies entirely within the representation power of the quantum feature space accessed via the kernel, not in speeding up the SVM optimization itself.
The QSVM algorithm operates as a hybrid quantum-classical procedure:
Let's break down the kernel computation step further. To estimate a single kernel entry kq(xi,xj)=∣⟨ϕ(xi)∣ϕ(xj)⟩∣2, we need a quantum circuit that measures the overlap or fidelity between the states ∣ϕ(xi)⟩=Uϕ(xi)∣0⟩⊗n and ∣ϕ(xj)⟩=Uϕ(xj)∣0⟩⊗n. Common techniques include:
Workflow for training a Quantum Support Vector Machine (QSVM). Quantum resources are used to estimate the kernel matrix entries, which are then processed by a classical SVM algorithm. Prediction also requires quantum kernel evaluations between the new data point and the support vectors.
Each estimation requires multiple measurements (shots) on the quantum computer to obtain a statistically reliable estimate of the probability. Calculating the full N×N kernel matrix for a training set of size N requires O(N2) calls to the quantum kernel estimation subroutine.
Recall the dual formulation of the classical soft-margin SVM for binary classification:
αmaxi=1∑Nαi−21i=1∑Nj=1∑Nαiαjyiyjk(xi,xj)subject to:
i=1∑Nαiyi=0 0≤αi≤C,for i=1,…,NHere, yi∈{−1,+1} are the class labels, αi are the Lagrange multipliers (non-zero only for support vectors), C is the regularization parameter, and k(xi,xj) is the kernel function.
In QSVM, we simply replace k(xi,xj) with the quantum kernel kq(xi,xj) estimated using one of the quantum circuit methods mentioned above. The optimization problem remains a convex quadratic program that is solved classically using standard numerical optimization libraries.
Once the optimal αi∗ and the bias b∗ are found, the prediction for a new data point xnew is given by:
ypred=sign(i=1∑Nαi∗yikq(xi,xnew)+b∗)Notice that prediction also requires computing quantum kernel entries, specifically between the new data point xnew and all the support vectors xi (those with αi∗>0).
QSVM represents a conceptually elegant way to integrate quantum feature spaces into a well-established machine learning framework. While practical advantages remain an active area of research, understanding its mechanism provides valuable insight into the potential and challenges of quantum kernel methods. The next section delves into the practical implementation details and challenges when applying QSVM to complex datasets.
© 2025 ApX Machine Learning