Ch.#8
Ch.#8
For the matrix in the intuitive opening example at the start of Sec. 8.1, the characteristic equation is
2
l ! 13l # 30 " (l ! 10)(l ! 3) " 0. The eigenvalues are {10, 3}. Corresponding eigenvectors are
[3 4]T and [!1 1]T , respectively. The reader may want to verify this. !
General Approach
This example illustrates the general case as follows. Equation (1) written in components is
Transferring the terms on the right side to the left side, we have
In matrix notation,
ByCHAP. 8 Linear
Cramer’s Algebra:
theorem in Matrix Eigenvalue
Sec. 7.7, Problems linear system of equations has a
this homogeneous
Cont.
nontrivial solution if and only if the corresponding determinant of the coefficients is zero:
By Cramer’s theorem in Sec. 7.7, this homogeneous linear system of equations has a
Á
nontrivial solution if and only if thea11 %l a12determinant
corresponding a1ncoefficients is zero:
of the
a21% l a22a%
a11 12
l Á Á aa1n
2n
(4) D(l) " det (A % lI) " 5 5 " 0.
#a #
a22 % l Á
Á #
a2n
21
(4) D(l) " det (A % lI) " 5 5 " 0.
an1 # an2# Á
Á ann #% l
HEOREM 1 The eigenvalues of a square matrix A are the roots of the characteristic equation
Eigenvalues
(4) of A.
The eigenvalues
Hence an n ! nof matrix
a square
hasmatrix A are
at least one the roots of the
eigenvalue andcharacteristic equation
at most n numerically
(4) of A.eigenvalues.
different
Hence an n ! n matrix has at least one eigenvalue and at most n numerically
different eigenvalues.
For larger n, the actual computation of eigenvalues will, in general, require the use
of Newton’s method (Sec. 19.2) or another numeric approximation method in Secs.
For larger n, the actual computation of eigenvalues will, in general, require the use
20.7–20.9.
of Newton’s method (Sec. 19.2) or another numeric approximation method in Secs.
The eigenvalues must be determined first. Once these are known, corresponding
Example 1 demonstrates how to systematically solve a simple eigenvalue problem.
"5 2
Example A! c d. A ! c
"5 2
d.
2 "2 2 "2
g the terms on the right to the left, we get ("5 " l)x 1 # 2x 2 !0
(2*)
2x 1 # ("2 " l)x 2 ! 0.
("5 " l)x 1 # 2x 2 !0
This can be written in matrix notation
2x 1 # ("2 " l)x 2 ! 0.
(3*) (A " lI)x ! 0
written in matrix notation
because (1) is Ax " lx ! Ax " lIx ! (A " lI)x ! 0, which gives (3*). We see that this is a homogeneous
linear system. By Cramer’s theorem in Sec. 7.7 it has a nontrivial solution x $ 0 (an eigenvector of A we are
(A " lI)x ! 0
looking for) if and only if its coefficient determinant is zero, that is,
"5 " l 2
) ! det (A " lI) ! 2 2 ! ("5 " l)("2 " l) " 4 ! l2 # 7l # 6 ! 0.
2 "2 " l
e Matrix Eigenvalue Problem. Determining Eigenvalues and Eigenvectors 325
Cont.
We call D (l) the characteristic determinant or, if expanded, the characteristic polynomial, and D (l) " 0
the characteristic equation of A. The solutions of this quadratic equation are l1 " !1 and l2 " !6. These
are the eigenvalues of A.
(b1) Eigenvector of A corresponding to l1. This vector is obtained from (2*) with l " l1 " !1, that is,
!4x 1 # 2x 2 " 0
2x 1 ! x 2 " 0.
A solution is x 2 " 2x 1, as we see from either of the two equations, so that we need only one of them. This
determines an eigenvector corresponding to l1 " !1 up to a scalar multiple. If we choose x 1 " 1, we obtain
the eigenvector
1 !5 2 1 !1
x1 " c d, Check: Ax1 " c dc d " c d " (!1)x1 " l1x1.
2 2 !2 2 !2
(b2) Eigenvector of A corresponding to l2. For l " l2 " !6, equation (2*) becomes
x 1 # 2x 2 " 0
2x 1 # 4x 2 " 0.
A solution is x 2 " !x 1>2 with arbitrary x1. If we choose x 1 " 2, we get x 2 " !1. Thus an eigenvector of A
A solution is x 2 " 2x 1, as we see from either of the two equations, so that we need only one of them. This
determines an eigenvector corresponding to l1 " !1 up to a scalar multiple. If we choose x 1 " 1, we obtain
the eigenvector
1 !5 2 1 !1
Cont.x1 " c d , Check: Ax1 " c dc d " c d " (!1)x1 " l1x1.
2 2 !2 2 !2
(b2) Eigenvector of A corresponding to l2. For l " l2 " !6, equation (2*) becomes
x 1 # 2x 2 " 0
2x 1 # 4x 2 " 0.
A solution is x 2 " !x 1>2 with arbitrary x1. If we choose x 1 " 2, we get x 2 " !1. Thus an eigenvector of A
corresponding to l2 " !6 is
2 !5 2 2 !12
x2 " c d, Check: Ax2 " c dc d " c d " (!6)x2 " l2x2.
!1 2 !2 !1 6
For the matrix in the intuitive opening example at the start of Sec. 8.1, the characteristic equation is
2
l ! 13l # 30 " (l ! 10)(l ! 3) " 0. The eigenvalues are {10, 3}. Corresponding eigenvectors are
[3 4]T and [!1 1]T , respectively. The reader may want to verify this. !
This example illustrates the general case as follows. Equation (1) written in components is
1 1
c d and c d. !
For larger n, the actual computation i of eigenvalues
!i will, in general, require the use
of Newton’s method (Sec. 19.2) or another numeric approximation method in Secs.
InCont.
20.7–20.9.
the next section we shall need the following simple theorem.
The eigenvalues must be determined first. Once these are known, corresponding
eigenvectors are obtained from the system (2), for instance, by the Gauss elimination,
EM 3 where l is the of
Eigenvalues eigenvalue for which an eigenvector is wanted. This is what we did in
the Transpose
Example 1 and shall
The transpose doa again
AT of squareinmatrix
the examples below.
A has the same (To prevent as
eigenvalues misunderstandings:
A.
numeric approximation methods, such as in Sec. 20.8, may determine eigenvectors first.)
Eigenvectors have the following properties.
OOF Transposition does not change the value of the characteristic determinant, as follows from
EM 2 Theorem 2d in Sec.
Eigenvectors, 7.7.
Eigenspace !
If w and x are eigenvectors of a matrix A corresponding to the same eigenvalue l,
Having
so aregained
w # xa (provided
first impression
x $ %wof )matrix
and kxeigenvalue
for any k $ problems,
0. we shall illustrate their
importance
Hence with some typicalcorresponding
the eigenvectors applications intoSec.
one8.2.
and the same eigenvalue l of A,
together with 0, form a vector space (cf. Sec. 7.4), called the eigenspace of A
corresponding to that l.
OOF Aw " lw and Ax " lx imply A(w # x) " Aw # Ax " lw # lx " l(w # x) and
A (kw) " k (Aw) " k (lw) " l (kw); hence A (kw # /x) " l (kw # /x). !
A!D 2 "6T . 2
1 "2 "3
A!D 2 1 "6T .
"1 "2 0
"1 "2 0
Solution. For our matrix, the characteristic determinant gives the characteristic equation
Solution. For our matrix, the characteristic determinant gives the characteristic equation
"l3 " l2 # 21l # 45 ! 0.
"l3 " l2 # 21l # 45 ! 0.
The roots (eigenvalues of A) are l1 ! 5, l2 ! l3 ! "3. (If you have trouble finding roots, you may want to
Thealgorithm
use a root finding roots (eigenvalues
such as of A) are lmethod
Newton’s 1 ! 5, l2(Sec.
! l319.2). (If you
! "3.Your CAShave
or trouble finding
scientific roots, you
calculator can may
find want to
use a root finding algorithm such as Newton’s method (Sec. 19.2). Your CAS or scientific calculator can find
roots. However, to really learn and remember this material, you have to do some exercises with paper and pencil.)
roots. However, to really learn and remember this material, you have to do some exercises with paper and pencil.)
To find eigenvectors, we apply the Gauss elimination (Sec. 7.3) to the system (A " lI)x ! 0, first with l ! 5
To find eigenvectors, we apply the Gauss elimination (Sec. 7.3) to the system (A " lI)x ! 0, first with l ! 5
and then with land "3. with
! then For ll ! 5 the characteristic matrix is
! "3. For l ! 5 the characteristic matrix is
"7 2 "3
"7 2 "3 "7 2
"7 "32 "3
A " lI ! A "
A 5I !!
" lI D A2 " "4
5I ! D"62T . "4 It"6
row-reduces
T. to D 0to " D24
It row-reduces
4824
7 0 ""7 7T . " 7 T .
48
4824 48
Hence
Hence it has rank 2. itChoosing
has rank x2.3 Choosing
! "1 wex 3have
! "1x 2 we
! 2have
from !24
x2 " 2 xfrom
7 2 "" 7 3x !
7 x 2 " x3 !
0 7and 0 and
then then
x1 ! x 1 ! 1 from
1 from
"7x 1 # 2x 2 ""7x
3x 31 !
# 0.2x 2Hence
" 3x 3an 0. Hence an of
!eigenvector eigenvector of A corresponding
A corresponding to l ! 5 istoxl1 ! [1 .2 "1]T.
! 5[1is x21 !"1] T
1 2 1
"3 2 "3 1 2 1"3 2 "3
A " lI ! A # 3I ! D 2 4 "6T row-reduces to D0 0 0T .
A " lI ! A # 3I ! D 2 4 "6T row-reduces to D0 0 0T .
"1 "2 3 0 0 0
"1 "2 3 0 0 0
"7 2 "3 "7 2 "3
A " lI ! A " 5I ! D 2 "4 "6T . It row-reduces to D 0 " 24
7 " 48
7 T.
Cont.
Hence it has rank 2. Choosing x 3 ! "1 we have x 2 ! 2 from " 24 48
7 x 2 " 7 x 3 ! 0 and then x 1 ! 1 from
"7x 1 # 2x 2 " 3x 3 ! 0. Hence an eigenvector of A corresponding to l ! 5 is x1 ! [1 2 "1]T.
For l ! "3 the characteristic matrix
1 2 "3 1 2 "3
A " lI ! A # 3I ! D 2 4 "6T row-reduces to D0 0 0T .
"1 "2 3 0 0 0
"2
x 2 ! D 1T
and
x3 ! D0T . !
1
and
Definitions x3 ! D0T . !
1
The
Algebra: Matrix order MProblems
Eigenvalue l of an eigenvalue l as a root of the characteristic polynomial is called the
algebraic multiplicity of l. The number m l of linearly independent eigenvectors
corresponding
racteristic polynomial has todegree
l is called geometric
the sum
n, the of all multiplicity of l. Thus m l is the dimension
the algebraic
of the eigenspace
st equal n. In Example 2 for l "corresponding
!3 we have mtol this
" Ml.l " 2. In general,
n be shown. The difference ¢ l " M l ! m l is called the defect of l.
n Example 2, but positive defects ¢ l can easily occur:
0 1 !l 1
A" c d is det (A ! lI) " 2 2 " l2 " 0.
0 0 0 !l
genvalue of algebraic multiplicity M 0 " 2. But its geometric multiplicity is only m 0 " 1,
ult from !0x 1 # x 2 " 0, hence x 2 " 0, in the form [x 1 0]T. Hence for l " 0 the defect
3 2 3!l 2
" c d is det (A ! lI) " 2 2 " (3 ! l)2 " 0.
0 3 0 3!l
CHAP. 8 Linear Algebra: Matrix Eigenvalue Problems
Since the characteristic polynomial has degree n, the sum of all the algebraic
multiplicities must equal n. In Example 2 for l " !3 we have m l " M l " 2. In general,
m l $ M l, as can be shown. The difference ¢ l " M l ! m l is called the defect of l.
Example
Thus ¢ !3 " 0 in Example 2, but positive defects ¢ l can easily occur:
0 1 !l 1
A" c d is det (A ! lI) " 2 2 " l2 " 0.
0 0 0 !l
Hence l " 0 is an eigenvalue of algebraic multiplicity M 0 " 2. But its geometric multiplicity is only m 0 " 1,
since eigenvectors result from !0x 1 # x 2 " 0, hence x 2 " 0, in the form [x 1 0]T. Hence for l " 0 the defect
is ¢ 0 " 1.
Similarly, the characteristic equation of the matrix
3 2 3!l 2
A" c d is det (A ! lI) " 2 2 " (3 ! l)2 " 0.
0 3 0 3!l
Hence l " 3 is an eigenvalue of algebraic multiplicity M 3 " 2, but its geometric multiplicity is only m 3 " 1,
since eigenvectors result from 0x 1 # 2x 2 " 0 in the form [x 1 0]T. !
Symmetric,andSkew-Symmetric,
Symmetric, Skew-Symmetric, Orthogonal Matrices and Orthogonal Matrices335
(3) AT ! A!1.
2 2 1
E 2 Illustration of Formula (4)
This basic theorem (and an extension of it) will be proved in Sec. 8.5.
y2 sin u cos u x2
b1
.
(7) a • b " aTb " [a1 Á an] D . T .
.
bn
That is, for any a and b in Rn, orthogonal n # n matrix A, and u " Aa, v " Ab
we have u • v " a • b.
Hence the transformation also preserves the length or norm of any vector a in
n
R given by
OOF Let A be orthogonal. Let u " Aa and v " Ab. We must show that u • v " a • b. Now
(Aa)T " aTAT by (10d) in Sec. 7.2 and ATA " A!1A " I by (3). Hence
(9) u • v " uTv " (Aa)TAb " aTATAb " aTIb " aTb " a • b.
Now Thethedeterminant
column vectors
of an of A (#A matrix
orthogonal ) are the
hasrow
thevectors of A.
value !1 Hence the row vectors
or "1.
of A also form an orthonormal system.
(b) Conversely, if the column vectors of A satisfy (10), the off-diagonal entries in (11)
must be 0 and the diagonal entries 1. Hence ATA # I, as (11) shows. Similarly, AAT # I.
PROOF This det AB
Fromimplies AT##det
A!1A because
det B (Sec.
also A7.8,
!1
ATheorem
# AA!1 4) # Iand AT # det
det inverse
and the A (Sec.Hence
is unique. 7.7,
Theorem
A 2d), we Similarly
is orthogonal. get for anwhen
orthogonal
the rowmatrix
vectors of A form an orthonormal system, by
Symmetric,
what has been saidSkew-Symmetric,
at the end of part (a).
!1
and Orthogonal
T T
Matrices
2
!
1 # det I # det (AA ) # det (AA ) # det A det A # (det A) . !
0:56 AM Page 337
AOMRPELME 4 Illustration of Theorems
Determinant 3 and 4 Matrix
of an Orthogonal
TheThe determinant
last matrix of 1anandorthogonal
in Example the matrix inmatrix has Theorems
(6) illustrate the value3 !1
and 4or
because
"1. their determinants are
"1 and !1, as you should verify. !
PROOF (a) Let A be orthogonal. Then A!1A # ATA # I. In terms of column vectors a1, Á , an,
This shows that we have decomposed the complicated action of A on an arbitrary vector
x into a sum of simple actions (multiplication by scalars) on the eigenvectors of A. This
is the point of an eigenbasis.
CHAP.Now8 ifLinear Algebra: Matrix
the n eigenvalues are allEigenvalue Problems
different, we do obtain a basis:
Eigenbases
CHAP. 8 Linear Algebra: Matrix Eigenvalue Problems
P EL M
R E 11 Eigenbasis. Nondistinct Eigenvalues. Nonexistence
Basis of Eigenvectors
MPLE 1 Eigenbasis. Nondistinct Eigenvalues. Nonexistence
5 3 A has n distinct eigenvalues,
If an n & n matrix 1 then1 A has a basis of eigenvectors
The matrix
Á A! c 5 3n d has a basis of eigenvectors c d ,1 c
1 d corresponding to the eigenvalues l1 ! 8,
x ,
The matrix
1 , x
An ! c 3 5d. has a basis of eigenvectors c d , c1 d #1
for R corresponding to the eigenvalues l1 ! 8,
3 5 1 #1
l2 ! 2. (See Example 1 in Sec. 8.2.)
l2 ! 2. (See Example 1 in Sec. 8.2.) n
Even if not all n eigenvalues are different, a matrix A may still provide an eigenbasis n for R . See Example 2
Even if not all n eigenvalues are different,Á , xna matrix A may still provide an eigenbasis for R . See Example 2
ROOF All
in we
Sec. have
8.1, to
where show
n
in Sec. 8.1, where n ! 3. ! is
3. that x 1, are linearly independent. Suppose they are not. Let
r beOn the
On the largest
the other hand,
other hand, AAmay
integer maysuch that
notnot have
have enough
Á , xlinearly
{xenough r } is
1, linearly a independent
linearlyeigenvectors
independent independent
eigenvectorsset.toup
to make Then
make rup%aFor
a basis. nbasis. For
instance,
and the A
instance, Asetinin{x 1,
ExampleÁ ,3 x3ofrof
Example xSec.
, Sec. }8.1
r!18.1 linearly dependent. Thus there are scalars c1, Á , cr!1,
is is
is
not all zero, such that
00 1 1 k k
AA!!c c dd andand
hashas
onlyonly
one one eigenvector c
eigenvector d c d(k $ 0(k
, arbitrary).
$ 0, arbitrary). ! !
(2) 00 0 0 c1x1 ! Á ! cr!1xr!1 # 00 0
Actually,
Actually, eigenbases
eigenbases exist
existunder much
under bymore
much general
Amore conditions
general # ljxthan
Axjconditions those in Theorem 1.
(see Sec. 7.4). Multiplying both sides and using j, we than those
obtain in Theorem 1.
An important case is the following.
An important case is the following.
and a more general result
(3) A(c1x1 ! Á ! cr!1xr!1) # c1l1x1 ! Á ! cr!1lr!1xr!1 # A0 # 0.
REM 2 Symmetric Matrices
EM 2 To Symmetric
get rid of theMatrices
last term, we subtract lr!1 times (2) from this, obtaining
A symmetric matrix has an orthonormal basis of eigenvectors for R n.
A symmetric matrixc1(l1has
$ lan orthonormal
)x ! Á ! c basis
(l $ of
l eigenvectors
)x # 0. for Rn.
r!1 1 r r r!1 r
For a proof (which is involved) see Ref. [B3], vol. 1, pp. 270–272.
Hereac1proof
(l1 $ (which Á , cr(lr $ lr!1) # 0 since {x 1, Á , x r } is linearly independent.
lr!1) #is0,involved)
For see Ref. [B3], vol. 1, pp. 270–272.
M P L E 2 Hence c1 # ÁBasis
Orthonormal # cof
r # 0, since all the eigenvalues are distinct. But with this, (2) reduces to
Eigenvectors
cr!1xr!1 # 0, hence cr!1 # 0, since xr!1 " 0 (an eigenvector!). This contradicts the factT
P L E 2 Orthonormal
The first matrix inBasis of Eigenvectors
Example 1 is symmetric, and an orthonormal basis of eigenvectors is 31> 12 1> 124 ,
Similarity of Matrices. Diagonalization
Similarity
Eigenbases of Matrices.
also play Diagonalization
a role in reducing a matrix A to a diagonal matrix whose entries are
the eigenvalues of A. This is done by a “similarity transformation,” which is defined as
Eigenbases
follows (andalso
willplay
havea role in reducing
various a matrix
applications A to a diagonal
in numerics in Chap.matrix
20). whose entries are
the eigenvalues of A. This is done by a “similarity transformation,” which is defined as
follows (and will have various applications in numerics in Chap. 20).
Similarity of Matrices and Diagonalization
NITION Similar Matrices. Similarity Transformation
NITION
An n " n matrix  is called similar to an n " n matrix A if
Similar Matrices. Similarity Transformation
An n " n matrix  is called similarÂto!anP !1
(4) n"APn matrix A if
!1
for some (nonsingular!) n " n matrix P. This AP
(4)  ! P transformation, which gives  from
A, is called a similarity transformation.
for some (nonsingular!) n " n matrix P. This transformation, which gives  from
A, is called a similarity transformation.
The key property of this transformation is that it preserves the eigenvalues of A:
The key property of this transformation is that it preserves the eigenvalues of A:
OREM 3 Eigenvalues and Eigenvectors of Similar Matrices
OREM 3 If  is similar
Eigenvalues A, then  has
andtoEigenvectors of the eigenvalues as A.
sameMatrices
Similar
Furthermore, if x is an eigenvector of A, then y ! P !1x is an eigenvector of Â
If  is similar to
corresponding A, then
to the sameÂeigenvalue.
has the same eigenvalues as A.
Furthermore, if x is an eigenvector of A, then y ! P !1x is an eigenvector of Â
corresponding to the same eigenvalue.
Indeed, these are eigenvectors of the diagonal matrix A.
Perhaps we see that x1 and x2 are the columns of P. This suggests the general method of transforming a
matrix A to diagonal form D by using P ! X, the matrix with eigenvectors as columns. !
M 4 Diagonalization of a Matrix
If an n " n matrix A has a basis of eigenvectors, then
(5) D ! X!1AX
is diagonal, with the eigenvalues of A as the entries on the main diagonal. Here X
is the matrix with these eigenvectors as column vectors. Also,
(5*) D m ! X!1AmX (m ! 2, 3, Á ).
CHAP. 8 Linear Algebra: Matrix Eigenvalue Problems
y1 5 3 x1 y1 " 5x 1 ! 3x 2
(1) y" c d " Ax " c d c d; in components,
y2 3 5 x2 y2 " 3x 1 ! 5x 2.
Find the principal directions, that is, the directions of the position vector x of P for which the direction of the
position vector y of Q is the same or exactly opposite. What shape does the boundary circle take under this
deformation?
Solution. We are looking for vectors x such that y " lx. Since y " Ax, this gives Ax " lx, the equation
of an eigenvalue problem. In components, Ax " lx is
5#l 3
(3) 2 2 " (5 # l)2 # 9 " 0.
3 5#l
y1 5 3 x1 y1 " 5x 1 ! 3x 2
(1) y" c d " Ax " c d c d; in components,
y2 3 5 x2 y2 " 3x 1 ! 5x 2.
Find the principal directions, that is, the directions of the position vector x of P for which the direction of the
position vector y of Q is the same or exactly opposite. What shape does the boundary circle take under this
Example
deformation?
Solution. We are looking for vectors x such that y " lx. Since y " Ax, this gives Ax " lx, the equation
of an eigenvalue problem. In components, Ax " lx is
5#l 3
(3) 2 2 " (5 # l)2 # 9 " 0.
3 5#l
Its solutions are l1 " 8 and l2 " 2. These are the eigenvalues of our problem. For l " l1 " 8, our system (2)
becomes
We thus obtain as eigenvectors of A, for instance, [1 1]T corresponding to l1 and [1 #1]T corresponding to
l2 (or a nonzero scalar multiple of these). These vectors make 45° and 135° angles with the positive x1-direction.
8.2 Some Applications
They give of theEigenvalue directions, the answer to our problem. The eigenvalues show that331
principal Problems in the principal
directions the membrane is stretched by factors 8 and 2, respectively; see Fig. 160.
Accordingly, if we choose the principal directions
x2 as directions of a new Cartesian u 1u 2-coordinate system,
say, with the positive u 1-semi-axis in the first quadrant and the positive u 2-semi-axis in the second quadrant of
the x 1x 2-system, and if we set u 1 " r cos !, u 2 " r sin !, then a boundary point of the unstretched circular
Pr irec
in ti
d
membrane has coordinates cos !, sin !. Hence, after the stretch we have
ci on
ct al
pa
re ip
n
l
io
di inc
Pr
Since cos2 ! ! sin2 ! " 1, this shows that the deformed boundary is an ellipse (Fig. 160)
z 21 z 22
(4) ! " 1. !
82 22
Fig. 160. Undeformed and deformed membrane in Example 1