Okay, you've successfully used the characteristic equation, det(A−λI)=0, to find the eigenvalues (λ) for a given square matrix A. Remember, eigenvalues represent the factors by which eigenvectors are scaled when transformed by the matrix A. Now, the next step is to find the actual eigenvectors (x) corresponding to each of these eigenvalues.
We start with the fundamental definition relating a matrix, its eigenvalue, and its corresponding eigenvector: Ax=λx
Our goal is to find the non-zero vector x that satisfies this equation for a specific eigenvalue λ. We can rearrange this equation to bring all terms to one side: Ax−λx=0
To proceed, we need to factor out the vector x. We can introduce the identity matrix I (of the same dimension as A) without changing the equation, as Ix=x: Ax−λIx=0
Now, we can factor out x: (A−λI)x=0
This equation is a homogeneous system of linear equations. It looks very similar to the form Mx=0, where M=(A−λI). We are looking for the non-zero solutions x to this system.
Why are we guaranteed to find non-zero solutions? Recall that we found the eigenvalues λ by solving det(A−λI)=0. A matrix with a determinant of zero is called a singular matrix. Singular matrices are not invertible, and a key property of the system Mx=0 when M is singular is that it has infinitely many non-trivial (non-zero) solutions. These non-zero solutions for x are precisely the eigenvectors associated with the eigenvalue λ.
So, for each eigenvalue λ you calculated:
Let's work through a simple 2x2 example. Consider the matrix: A=[4213]
Suppose we have already found the eigenvalues using the characteristic equation. Let's say one of the eigenvalues is λ=5. Now we find the eigenvector(s) corresponding to λ=5.
Form the Matrix (A−λI): A−λI=A−5I=[4213]−5[1001]=[4213]−[5005]=[−121−2]
Set up the System (A−λI)x=0: Let x=[x1x2]. The system is: [−121−2][x1x2]=[00] This translates to the equations: −x1+x2=0 2x1−2x2=0
Solve the System: Notice that the second equation is just -2 times the first equation. They are linearly dependent, which is expected since the matrix (A−5I) is singular. From the first equation, we get: x2=x1 The variable x1 is a free variable. We can choose any value for it (except 0, because eigenvectors must be non-zero), and x2 will be determined.
Identify an Eigenvector: Let's choose a simple value for the free variable, say x1=1. Then x2=1. So, one eigenvector corresponding to the eigenvalue λ=5 is: x=[11] Any non-zero scalar multiple of this vector, like [22] or [−0.5−0.5], is also a valid eigenvector for λ=5. The set of all such vectors forms the eigenspace for λ=5.
You would repeat this process for any other eigenvalues found for matrix A.
While performing Gaussian elimination by hand is instructive for smaller matrices, for larger matrices encountered in machine learning applications, you'll rely on numerical libraries like NumPy, which have efficient functions to compute both eigenvalues and eigenvectors directly. Understanding this underlying process, however, provides valuable insight into what these functions are doing and how to interpret their results.
© 2025 ApX Machine Learning