Title: 7'1 Eigenvalues And Eigenvectors
17.1 Eigenvalues And Eigenvectors
2Definition
- If A is an nn matrix, then a nonzero vector x in
Rn is called an eigenvector of A if Ax is a
scalar multiple of x that is, - Ax?x
- for some scalar ?. The scalar ? is called an
eigenvalue of A, and x is said to be an
eigenvector of A corresponding to ?.
3Example 1Eigenvector of a 22 Matrix
- The vector is an eigenvector of
- Corresponding to the eigenvalue ?3, since
4- To find the eigenvalues of an nn matrix A we
rewrite Ax?x as - Ax?Ix
- or equivalently,
- (?I-A)x0
(1) - For ? to be an eigenvalue, there must be a
nonzero solution of this equation. However, by
Theorem 6.4.5, Equation (1) has a nonzero
solution if and only if - det (?I-A)0
- This is called the characteristic equation of A
the scalar satisfying this equation are the
eigenvalues of A. When expanded, the determinant
det (?I-A) is a polynomial p in ? called the
characteristic polynomial of A.
5Example 2Eigenvalues of a 33 Matrix (1/3)
- Find the eigenvalues of
- Solution.
- The characteristic polynomial of A is
- The eigenvalues of A must therefore satisfy the
cubic equation
6Example 2Eigenvalues of a 33 Matrix (2/3)
- To solve this equation, we shall begin by
searching for integer solutions. This task can be
greatly simplified by exploiting the fact that
all integer solutions (if there are any) to a
polynomial equation with integer coefficients - ?nc1?n-1cn0
- must be divisors of the constant term cn. Thus,
the only possible integer solutions of (2) are
the divisors of -4, that is, 1, 2, 4.
Successively substituting these values in (2)
shows that ?4 is an integer solution. As a
consequence, ?-4 must be a factor of the left
side of (2). Dividing ?-4 into ?3-8?217?-4 show
that (2) can be rewritten as - (?-4)(?2-4?1)0
7Example 2Eigenvalues of a 33 Matrix (3/3)
- Thus, the remaining solutions of (2) satisfy the
quadratic equation - ?2-4?10
- which can be solved by the quadratic formula.
Thus, the eigenvalues of A are
8Example 3Eigenvalues of an Upper Triangular
Matrix (1/2)
- Find the eigenvalues of the upper triangular
matrix - Solution.
- Recalling that the determinant of a triangular
matrix is the product of the entries on the main
diagonal (Theorem 2.2.2), we obtain
9Example 3Eigenvalues of an Upper Triangular
Matrix (2/2)
- Thus, the characteristic equation is
- (?-a11)(?-a22) (?-a33) (?-a44)0
- and the eigenvalues are
- ?a11, ?a22, ?a33, ?a44
- which are precisely the diagonal entries of A.
10Theorem 7.1.1
- If A is an nn triangular matrix (upper
triangular, low triangular, or diagonal), then
the eigenvalues of A are entries on the main
diagonal of A.
11Example 4Eigenvalues of a Lower Triangular Matrix
- By inspection, the eigenvalues of the lower
triangular matrix - are ?1/2, ?2/3, and ?-1/4.
12Theorem 7.1.2Equivalent Statements
- If A is an nn matrix and ? is a real number,
then the following are equivalent. - ? is an eigenvalue of A.
- The system of equations (?I-A)x0 has nontrivial
solutions. - There is a nonzero vector x in Rn such that
Ax?x. - ? is a solution of the characteristic equation
det(?I-A)0.
13Finding Bases for Eigenspaces
- The eigenvectors of A corresponding to an
eigenvalue ? are the nonzero x that satisfy
Ax?x. Equivalently, the eigenvectors
corresponding to ? are the nonzero vectors in the
solution space of (?I-A)x0. We call this
solution space the eigenspace of A corresponding
to ?.
14Example 5Bases for Eigenspaces (1/5)
- Find bases for the eigenspaces of
- Solution.
- The characteristic equation of matrix A is
?3-5?28?-40, or in factored form,
(?-1)(?-2)20 thus, the eigenvalues of A are ?1
and ?2, so there are two eigenspaces of A.
15Example 5Bases for Eigenspaces (2/5)
- By definition,
- Is an eigenvector of A corresponding to ? if and
only if x is a nontrivial solution of (?I-A)x0,
that is, of - If ?2, then (3) becomes
16Example 5Bases for Eigenspaces (3/5)
- Solving this system yield
- x1-s, x2t, x3s
- Thus, the eigenvectors of A corresponding to ?2
are the nonzero vectors of the form - Since
17Example 5Bases for Eigenspaces (4/5)
- are linearly independent, these vectors form a
basis for the eigenspace corresponding to ?2. - If ?1, then (3) becomes
- Solving this system yields
- x1-2s, x2s, x3s
18Example 5Bases for Eigenspaces (5/5)
- Thus, the eigenvectors corresponding to ?1 are
the nonzero vectors of the form - is a basis for the eigenspace corresponding to
?1.
19Theorem 7.1.3
- If k is a positive integer, ? is an eigenvalue of
a matrix A, and x is corresponding eigenvector,
then ?k is an eigenvalue of Ak and x is a
corresponding eigenvector.
20Example 6Using Theorem 7.1.3 (1/2)
- In Example 5 we showed that the eigenvalues of
- are ?2 and ?1, so from Theorem 7.1.3 both
?27128 and ?171 are eigenvalues of A7. We
also showed that - are eigenvectors of A corresponding to the
eigenvalue ?2, so from Theorem 7.1.3 they are
also eigenvectors of A7 corresponding to
?27128. Similarly, the eigenvector
21Example 6Using Theorem 7.1.3 (2/2)
- of A corresponding to the eigenvalue ?1 is also
eigenvector of A7 corresponding to ?171.
22Theorem 7.1.4
- A square matrix A is invertible if and only if
?0 is not an eigenvalue of A.
23Example 7Using Theorem 7.1.4
- The matrix A in Example 5 is invertible since it
has eigenvalues ?1 and ?2, neither of which is
zero. We leave it for reader to check this
conclusion by showing that det(A)?0
24Theorem 7.1.5Equivalent Statements (1/3)
- If A is an nn matrix, and if TA Rn ?Rn is
multiplication by A, then the following are
equivalent. - A is invertible.
- Ax0 has only the trivial solution.
- The reduced row-echelon form of A is In.
- A is expressible as a product of elementary
matrix. - Axb is consistent for every n1 matrix b.
- Axb has exactly one solution for every n1
matrix b. - det(A)?0.
25Theorem 7.1.5Equivalent Statements (2/3)
- The range of TA is Rn.
- TA is one-to-one.
- The column vectors of A are linearly independent.
- The row vectors of A are linearly independent.
- The column vectors of A span Rn.
- The row vectors of A span Rn.
- The column vectors of A form a basis for Rn.
- The row vectors of A form a basis for Rn.
26Theorem 7.1.5Equivalent Statements (3/3)
- A has rank n.
- A has nullity 0.
- The orthogonal complement of the nullspace of A
is Rn. - The orthogonal complement of the row space of A
is 0. - ATA is invertible.
- ?0 is not eigenvalue of A.
277.2 Diagonalization
28Definition
- A square matrix A is called diagonalizable if
there is an invertible matrix P such that P-1AP
is a diagonal matrix the matrix P is said to
diagonalize A.
29Theorem 7.2.1
- If A is an nn matrix, then the following are
equivalent. - A is diagonalizable.
- A has n linearly independent eigenvectors.
30Procedure for Diagonalizing a Matrix
- The preceding theorem guarantees that an nn
matrix A with n linearly independent eigenvectors
is diagonalizable, and the proof provides the
following method for diagonalizing A. - Step 1. Find n linear independent eigenvectors of
A, say, p1, p2, , pn. - Step 2. From the matrix P having p1, p2, , pn as
its column vectors. - Step 3. The matrix P-1AP will then be diagonal
with ?1, ?2, , ?n as its successive diagonal
entries, where ?i is the eigenvalue corresponding
to pi, for i1, 2, , n.
31Example 1Finding a Matrix P That Diagonalizes a
Matrix A (1/2)
- Find a matrix P that diagonalizes
- Solution.
- From Example 5 of the preceding section we found
the characteristic equation of A to be - (?-1)(?-2)20
- and we found the following bases for the
eigenspaces
32Example 1Finding a Matrix P That Diagonalizes a
Matrix A (2/2)
- There are three basis vectors in total, so the
matrix A is diagonalizable and - diagonalizes A. As a check, the reader should
verify that
33Example 2A Matrix That Is Not Diagonalizable
(1/4)
- Find a matrix P that diagonalize
- Solution.
- The characteristic polynomial of A is
34Example 2A Matrix That Is Not Diagonalizable
(2/4)
- so the characteristic equation is
- (?-1)(?-2)20
- Thus, the eigenvalues of A are ?1 and ?2. We
leave it for the reader to show that bases for
the eigenspaces are - Since A is a 33 matrix and there are only two
basis vectors in total, A is not diagonalizable.
35Example 2A Matrix That Is Not Diagonalizable
(3/4)
- Alternative Solution.
- If one is interested only in determining whether
a matrix is diagonalizable and is not concerned
with actually finding a diagonalizing matrix P,
then it is not necessary to compute bases for the
eigenspaces it suffices to find the dimensions
of the eigenspaces. For this example, the
eigenspace corresponding to ?1 is the solution
space of the system - The coefficient matrix has rank 2. Thus, the
nullity of this matrix is 1 by Theorem 5.6.3, and
hence the solution space is one-dimensional.
36Example 2A Matrix That Is Not Diagonalizable
(4/4)
- The eigenspace corresponding to ?2 is the
solution space system - This coefficient matrix also has rank 2 and
nullity 1, so the eigenspace corresponding to ?2
is also one-dimensional. Since the eigenspaces
produce a total of two basis vectors, the matrix
A is not diagonalizable.
37Theorem 7.2.2
- If v1, v2, vk, are eigenvectors of A
corresponding to distinct eigenvalues ?1, ?2, ,
?k, thenv1, v2, vk is a linearly independent
set.
38Theorem 7.2.3
- If an nn matrix A has n distinct eigenvalues,
then A is diagonalizable.
39Example 3Using Theorem 7.2.3
- We saw in Example 2 of the preceding section that
- has three distinct eigenvalues,
. Therefore, A is
diagonalizable. Further, - for some invertible matrix P. If desired, the
matrix P can be found using method shown in
Example 1 of this section.
40Example 4A Diagonalizable Matrix
- From Theorem 7.1.1 the eigenvalues of a
triangular matrix are the entries on its main
diagonal. This, a triangular matrix with distinct
entries on the main diagonal is diagonalizable.
For example, - is a diagonalizable matrix.
41Theorem 7.2.4Geometric and Algebraic Multiplicity
- If A is a square matrix, then
- For every eigenvalue of A the geometric
multiplicity is less than or equal to the
algebraic multiplicity. - A is diagonalizable if and only if the geometric
multiplicity is equal to the algebraic
multiplicity for every eigenvalue.
42Computing Powers of a Matrix (1/2)
- There are numerous problems in applied
mathematics that require the computation of high
powers of a square matrix. We shall conclude this
section by showing how diagonalization can be
used to simplify such computations for
diagonalizable matrices. - If A is an nn matrix and P is an invertible
matrix, then - (P-1AP)2P-1APP-1APP-1AIAPP-1A2P
- More generally, for any positive integer k
- (P-1AP)kP-1AkP
(8)
43Computing Powers of a Matrix (2/2)
- It follows form this equation that if A is
diagonalizable, and P-1APD is a diagonal matrix,
then - P-1AkP(P-1AP)kDk
(9) - Solving this equation for Ak yields
- AkPDkP-1
(10) - This last equation expresses the kth power of A
in terms of the kth power of the diagonal matrix
D. But Dk is easy to compute for example, if -
44Example 5 Power of a Matrix (1/2)
- Using (10) to find A13, where
- Solution.
- We showed in Example 1 that the matrix A is
diagonalized by - and that
45Example 5 Power of a Matrix (2/2)
467.3 Orthogonal Diagonalization
47The Orthogonal Diagonalization Matrix Form
- Given an nn matrix A, if there exist an
orthogonal matrix P such that the matrix
P-1APPTAP, then A is said to be orthogonally
diagonalizable and P is said to orthogonally
diagonalize A.
48Theorem 7.3.1
- If A is an nn matrix, then the following are
equivalent. - A is orthogonally diagonalizable.
- A has an orthonormal set of n eigenvectors.
- A is symmetric.
49Theorem 7.3.2
- If A is a symmetric matrix, then
- The eigenvalues of A are real numbers.
- Eigenvectors from different eigenspaces are
orthogonal.
50Diagonalization of Symmetric Matrices
- As a consequence of the preceding theorem we
obtain the following procedure for orthogonally
diagonalizing a symmetric matrix. - Step 1. Find a basis for each eigenspace of A.
- Step 2. Apply the Gram-Schmidt process to each of
these bases to obtain an orthonormal basis for
each eigenspace. - Step 3. Form the matrix P whose columns are the
basis vectors constructed in Step2 this matrix
orthogonally diagonalizes A.
51Example 1An Orthogonal Matrix P That
Diagonalizes a Matrix A (1/3)
- Find an orthogonal matrix P that diagonalizes
- Solution.
- The characteristic equation of A is
52Example 1An Orthogonal Matrix P That
Diagonalizes a Matrix A (2/3)
- Thus, the eigenvalues of A are ?2 and ?8. By
the method used in Example 5 of Section 7.1, it
can be shown that - form a basis for the eigenspace corresponding to
?2. Applying the Gram-Schmidt process to u1,
u2 yields the following orthonormal
eigenvectors
53Example 1An Orthogonal Matrix P That
Diagonalizes a Matrix A (3/3)
- The eigenspace corresponding to ?8 has
- as a basis. Applying the Gram-Schmidt process to
u3 yields - Finally, using v1, v2, and v3 as column vectors
we obtain - which orthogonally diagonalizes A.