• Given a linear map from to , we want to find the directions where the inputs do not change
    • Say that we have some subspace , such as plane or line. If , this means that all can do is scale at most
    • This subspace is called an invariant subspace with regards to

Definition: Eigenvalues Let be a linear transformation. Let , and is the span of . If is invariant under , then that means for some , that . is called an eigenvalue of and is an associated eigenvector.

  • Essentially, for all of the eigenvectors, the transform can be simplified to multiplication by scalar
  • Here is an alternate definition for eigenvalues and eigenvectors, with respect to matrices

Definition: Let . is called an eigenvalue of if there exists such that , and is called the associated eigenvector. is called an eigenvalue of if . That is, if solves

Finding eigenvalues

  • Now, how do we actually end up finding eigenvalues?
    • You have to start with setting the transformation and scalar multiplication

We know that the definition of an eigenvalues is that for some linear transformation , that Rearranging the equation, we can write Hence, is not an isomorphism (since it is not one to one). The matrix associated: is not invertible, meaning that This gives us the characteristic polynomial for . The solutions to this polynomial are the eigenvalues of .

Definition: Characteristic polynomial. For some transformation , the characteristic polynomial is:

The values of that solve this equation are the eigenvalues () of .

Example: Find the eigenvalues of Eigenvalues are zeros of The solutions to this equation are the eigenvalues of . In this case, there are no real solutions, so the eigenvalues are complex.

Remark: Let be an eigenvalue of and is an eigenvector associated with . Then,

  • This remark shows that any eigenvalue will have an infinite number of associated eigenvectors
    • This is because due to linearity, we can multiply the vector by any real number

If is an eigenvector of , then is also an eigenvector of , for all Suppose is linearly independent of eigenvectors of . Then all elements of is an eigenvector of . Proof: TBA

Eigenvalues and eigenvector for 3x3 matrices

Example: Find the eigenvalues of Notice that when , the entire column is equal to zero, and therefore the determinant is equal to zero. This means that is one of the eigenvalues of this matrix. The same follows for .

  • To find the eigenvector for a given eigenvalue , we have to solve for the system of equations where

Example: Find the eigenvectors for in the previous problem. Our eigenvector is ,where , as . This implies that . We have to solve . This gives us that , , is a free variable, so we can say that is an example of an eigenvector of .