While the Singular Value Decomposition (SVD) formula, , might look abstract, it describes a fundamental geometric process. Understanding this geometry provides intuition for why SVD is so effective in applications like dimensionality reduction and understanding linear transformations.
Consider a linear transformation represented by an matrix . When we apply this transformation to a vector to get , SVD tells us this transformation can be broken down into three distinct geometric steps:
Rotation/Reflection (): The matrix (the transpose of ) is an orthogonal matrix. When applied to the input vector , performs a rotation and possibly a reflection of the input space . It doesn't change the lengths of vectors or the angles between them. Think of this step as aligning the input space along a special set of orthogonal directions, given by the columns of (the right singular vectors). These vectors form an orthonormal basis for the input space. essentially rotates the space so that these principal input directions align with the standard coordinate axes.
Scaling (): The matrix is an rectangular diagonal matrix. Its diagonal entries are the singular values (where is the rank of ), and all other entries are zero. This matrix scales the coordinates of the rotated vector . Specifically, it scales the -th coordinate (which corresponds to the direction of ) by the singular value . Directions corresponding to zero singular values are effectively squashed to zero. This step stretches or shrinks the space along the newly aligned axes.
Rotation/Reflection (): The matrix is also an orthogonal matrix (). This final step takes the scaled vector and performs another rotation and possibly a reflection in the output space . The columns of , (the left singular vectors), form an orthonormal basis for the output space. This step rotates the scaled vectors from the axes-aligned orientation (after step 2) into their final positions in the output space, aligning them with the principal output directions defined by the columns of .
Imagine applying the transformation to all the points on a unit circle in 2D (or a unit sphere in 3D). Here's how SVD breaks down the transformation geometrically:
The SVD essentially tells us that any linear transformation maps orthonormal basis vectors in the input space (columns of ) to orthogonal vectors in the output space (columns of scaled by singular values ). That is, for .
The transformation defined by matrix maps the unit circle (gray) to an ellipse (red). SVD decomposes this overall transformation into a sequence of rotation (), axis-aligned scaling (), and another rotation ().
This geometric view is particularly insightful for understanding dimensionality reduction using SVD. The singular values quantify the importance of each principal direction. Larger singular values correspond to directions where the data (or the transformation) has the most variance or "spread". By keeping only the components corresponding to the largest singular values, we retain the most significant geometric features of the transformation while potentially discarding dimensions associated with small singular values (which might represent noise or less important variations).
Cleaner syntax. Built-in debugging. Production-ready from day one.
Built for the AI systems behind ApX Machine Learning
Was this section helpful?
© 2026 ApX Machine LearningEngineered with