Eigenvalues and Eigenspaces

(insert picture of what a eigenvector is)

Definition: Let A be a n × n matrix. Then a nonzero vector u is an eigenvector if there exists a scalar λ such that Au = λu. The scalar λ here is called the eigenvalue. Here u is an eigenvector associated to λ.

Examples:

Theorem A square matrix is invertible if and only if 0 is not a eigenvalue.

Theorem/Definition: Let A be a n × n matrix with eigenvalue λ. Then the set of all eigenvectors associated to λ along with 0 forms a subspace, called the eigenspace, of n. This is also the null space of A − λI.

Theorem/Definition: Let A be an n × n matrix. Then λ is an eigenvalue if and only if det (A − λI) = 0. The polynomial det (A − λI) is called the charateristic polynomial of A. The multiplicity of a eigenvalue is its multiplicity in the charateristic polynomial.

Example: Find the eigenvalues and a basis for each eigenspace for A = [[0, 2,  − 1], [1,  − 1, 0], [1,  − 2, 0]].

It turns out that det (A − λI) is  − λ3 − λ2 + λ + 1 =  − (λ − 1)(λ + 1)2.

So we are just finding the basis for the nullspaces of A − I and A + I which we can do with row reductions.

Theorem: Let A be a square matrix with eigenvalue λ. Then the dimension of the associated eigenspace is less than or equal to the multiplicty of λ.