While Graph Convolutional Networks (GCNs) provide a powerful link between spectral graph theory and practical deep learning on graphs, their underlying spectral filter is relatively simple, often a first-order polynomial of the graph Laplacian. This limitation restricts their ability to capture complex spectral patterns and learn highly localized filters. To address this, more sophisticated spectral GNN architectures have been developed, primarily focusing on designing more expressive and flexible graph filters gθ(Λ). Two significant examples are ChebNet, which uses Chebyshev polynomial approximations, and CayleyNets, which employ rational Cayley filters.
The core idea behind ChebNet [Defferrard et al., 2016] is to approximate the desired spectral filter gθ(Λ) using a basis of orthogonal polynomials, specifically Chebyshev polynomials of the first kind, Tk(x). These polynomials are defined by the recurrence relation:
Tk(x)=2xTk−1(x)−Tk−2(x)with base cases T0(x)=1 and T1(x)=x.
Chebyshev polynomials are particularly suitable because they are defined on the interval [−1,1], and the eigenvalues of the normalized graph Laplacian can be scaled to fit this range. Let L be the graph Laplacian (e.g., L=I−D−1/2AD−1/2). Its eigenvalues λi lie in [0,λmax], where λmax is the largest eigenvalue (often λmax≤2 for the normalized Laplacian). We define the scaled Laplacian L~=λmax2L−I, whose eigenvalues λ~i fall within [−1,1].
The spectral filter gθ(Λ) acting on the diagonal matrix of eigenvalues Λ is then approximated by a K-th order polynomial:
gθ(Λ)≈k=0∑KθkTk(Λ~)Here, θ=(θ0,...,θK) are the learnable filter parameters (coefficients of the polynomial basis), and Tk(Λ~) applies the k-th Chebyshev polynomial function to each eigenvalue on the diagonal of Λ~.
Applying this filter to a graph signal x gives:
y=gθ(L)x=Ugθ(Λ)UTx≈U(k=0∑KθkTk(Λ~))UTx=k=0∑Kθk(UTk(Λ~)UT)xCrucially, UTk(Λ~)UT=Tk(UΛ~UT)=Tk(L~). This means we can compute the filtered signal without explicit eigendecomposition:
y≈k=0∑KθkTk(L~)xThe term Tk(L~)x can be computed efficiently using the Chebyshev recurrence relation directly on the scaled Laplacian L~ and the signal x. A fundamental property is that applying Tk(L~) corresponds to aggregating information strictly from the k-hop neighborhood of each node. This makes the filter strictly localized within K hops.
A ChebNet layer with input features H(l)∈RN×Fin and output features H(l+1)∈RN×Fout is defined as:
H:,j(l+1)=σ(i=1∑Fink=0∑KΘk,i,j(l)Tk(L~)H:,i(l))for j=1,...,Foutor more compactly, viewing Θk(l) as parameter matrices of size Fin×Fout:
H(l+1)=σ(k=0∑KTk(L~)H(l)Θk(l))where σ is a non-linear activation function.
Relationship to GCN: GCN can be viewed as a simplification of ChebNet where K=1. Setting K=1 gives y≈θ0T0(L~)x+θ1T1(L~)x=θ0x+θ1L~x. By further assuming λmax≈2, so L~≈L−I, and using a single parameter θ=θ0=−θ1, we recover an operation similar to the GCN propagation rule (after renormalization tricks). ChebNet, with K>1, allows for learning much more complex, localized filters by controlling the polynomial order K and the coefficients θk.
CayleyNets [Levie et al., 2018] propose using rational spectral filters based on the Cayley transform, aiming for better localization in the spectral domain, particularly for selecting specific frequency bands (spectral zooming).
The Cayley transform is defined as c(x,h)=(x−ih)/(x+ih), where h>0 is a real parameter and i is the imaginary unit. Cayley polynomials are built upon this transform. A basic Cayley filter acting on the Laplacian L is defined as:
gθ(L)=θ0I+2Rej=1∑rθj(hL−iI)j(hL+iI)−jHere, r controls the filter's complexity (similar to K in ChebNet), and θ=(θ0,...,θr) are the learnable coefficients (potentially complex, but often constrained). The parameter h acts as a spectral zoom parameter. Small h focuses the filter's energy near eigenvalue λ=0 (low frequencies), while large h distributes it more broadly.
Applying this filter requires computing terms like (hL+iI)−jx. This involves solving complex linear systems of the form (hL+iI)y=z. Since hL+iI is related to the Laplacian, these systems are typically sparse and can be solved efficiently using iterative methods like conjugate gradient.
Advantages: Cayley filters, being rational functions, can approximate band-pass filters more effectively than polynomials of the same order. This allows CayleyNets to potentially isolate specific frequency ranges in the graph signal, which might be beneficial for certain tasks where distinguishing between low-frequency (smooth) and high-frequency (variant) components is important.
Feature | ChebNet | CayleyNets |
---|---|---|
Filter Type | Polynomial (Chebyshev) | Rational (Cayley Transform) |
Localization | Spatial: Strictly K-hop neighborhood | Spectral: Better frequency band selection |
Parameters | Polynomial order K, coefficients θk | Filter order r, zoom parameter h, coefficients θj |
Computation | Recursive polynomial application (Tk(L)x) | Solving complex linear systems ((hL+iI)y=z) |
Flexibility | Good general-purpose filter approximation | Better spectral zooming capabilities |
Complexity | $O(K | \mathcal{E} |
Both ChebNet and CayleyNets represent significant advancements over the basic GCN filter by enabling more complex and tailored spectral filtering. ChebNet provides spatially localized filters whose reach (K) is explicitly controlled. CayleyNets offer potentially sharper filters in the frequency domain. The choice between them depends on the specific application requirements: if explicit control over spatial neighborhood size is desired, ChebNet is a natural choice. If precise frequency selection is the goal, CayleyNets might be more suitable.
These advanced spectral methods demonstrate the richness of designing GNNs from a graph signal processing perspective. While spatial methods (discussed next) have become very popular due to their inductive capabilities and simpler formulation, understanding spectral approaches like ChebNet and CayleyNets provides valuable insights into the expressive power and theoretical underpinnings of graph convolutions. Implementations of layers like ChebConv
are available in standard GNN libraries, allowing for practical experimentation with these powerful architectures.
© 2025 ApX Machine Learning