Canonical Forms Linear Algebra Notes: Satya Mandal October 25, 2005
Canonical Forms Linear Algebra Notes: Satya Mandal October 25, 2005
1 Introduction
Here F will denote a field and V will denote a vector space of dimen-
sion dim(V ) = n. (In this note, unless otherwise stated, n = dim(V ))
We will study operatores T on V. The goal is to investigate if we
can find a basis e1 , . . . , en such that
T (ei ) = λi ei .
2 Characteristic Values
2.1 Basic Definitions and Facts
Here we will discuss basic facts.
2.1 (Definition.) Let V be a vector space over a field F and T ∈
L(V, V ) be linear operator.
1
2. This non-zero element e ∈ V above is called a characterisitic
vector of T associated to λ. A characterisitic vector is also
known an eigen vector.
3. Write
N (λ) = {v ∈ V : T (v) = λv}.
Then N (λ) is a subspace of V and is said to be the charac-
terisitic space or eigen space of T associated to λ.
2.2 (Lemma.) Let V be a vector space over a field F and T ∈
L(V, V ) Then T is singular if and only if det(T ) = 0.
Proof. (⇒): We have T (e1 ) = 0 for some e1 ∈ V with e1 6= 0. We
can extend e1 to a basis e1 , e2 , . . . , en of V. Let A be the matrix of
T with respect to this basis. Since, T (e1 ) = 0, the first row of A is
zero. Therefore,
det(T ) = det(A) = 0.
So, this implication is established.
(⇐): Suppose det(T ) = 0. Let e1 , e2 , . . . , en be a basis of V and
A be the matrix of T with respect to this basis. So
det(T ) = det(A) = 0.
Therefore, A is not invertible. Hence AX = 0 for some non-zero
column
c1
c2
X= ... .
cn
Write v = c1 e1 + c2 e2 + · · · + cn en . Since not all ci is zero, v 6= 0.
Also, X
T (v) = ci T (ei )
c1 c1
c2 c2
. . . = (e1 , e2 , . . . , en )A . . . = 0..
= (T (e1 ), T (e2 ), . . . , T (en ))
cn cn
So, T is singular.
2
2.3 (Theorem.) Let V be a vector space over a field F and T ∈
L(V, V ) be linear operator. Let λ ∈ F be a scalar. Then the following
are equivalent:
1. λ is a characteristic value of T.
3. det(T − λI) = 0.
3
1. Let A be the matrix of T with respect to some basis E of
V. We define the characteristic polynomial of T to be the
characteristic polynomial of A. Note that this polynomial is
well defined by the above lemma.
2. We say that T is diagonalizable if the there is a basis e1 , . . . , en
such that each ei is a characteristic value of T. In this case,
T (ei ) = λi ei for some λi ∈ F. Hence, with respect to this basis,
the matrix of T = Diagonal(λ1 , λ2 , . . . , λn )
Depending on how many of these eigen values λi are distinct,
we can rewrite the matrix of T.
3. Also note, if T is diagonilazable as above, the characteric poly-
nomial of T = (X − λ1 )(X − λ2 ) · · · (X − λn ), which is com-
pletely factorizable.
4. Suppose T is diagonalizable, as above. Depending on how
many of these eigen values λi are distinct, we can rewrite the
matrix of T.
Now sSuppose T is diagonalizable c1 , c2 , . . . , cr are the distinct
eigen values of T. Then the matrix of T with respect to some
basis of V looks like:
c 1 Id 1 0 ··· 0
0 c 2 Id 2 · · · 0
··· ··· ··· ···
0 0 · · · c r Id r
where Ik is the identity matrix of order k. So, d1 +d2 +· · ·+dr =
n = dim(V ).
In this case, the characteristic polynomial of
Further,
di = dim(N (ci )).
(see 3 of 2.1.)
4
2.2 Decomposition of V
2.8 (Definition) Suppose V is vector space of a field F with dim(V ) =
n. Let f (X) = a0 + a1 X + a2 X 2 + · · · + ar X r ∈ F[X] be polynomial
and T ∈ L(V, V ) be linear operator. Then, by definition,
f (T ) = a0 Id + a1 T + a2 T 2 + · · · + ar T r ∈ L(V, V )
is an operator. So, L(V, V ) becomes a module over F[X].
2.9 (Remark) Suppose V is vector space of a field F with dim(V ) =
n. Let T ∈ L(V, V ) be linear operator. Let f (X) be a characteristic
polynomial of T. We have understanable interest how f (T ) works.
2.10 (Lemma) Suppose V is vector space of a field F with dim(V ) =
n. Let T ∈ L(V, V ) be linear operator. Let f (X) ∈ F[X] be any
polynomial. Suppose
T (v) = λv
for some v ∈ V and λ ∈ F. Then
f (T )(v) = f (λ)v.
The proof is obvious. This means if λ is an eigen value of T then
f (λ) is an eigen value of f (T )
2.11 (Lemma) Suppose V is vector space of a field F with dim(V ) =
n. Let T ∈ L(V, V ) be linear operator. Suppose c1 , . . . , ck are the
distinct eigen values of T. Let
Wi = N (ci )
be the eigen space of T associated to ci . Write
W = W 1 + W2 + · · · + W k .
Then
dim(W ) = dim(W1 ) + dim(W2 ) + · · · + dim(Wk ).
Indeed, if
Ei = {eij ∈ Wi : j = 1, . . . , di }
is a basis of Wi , then
E = {eij ∈ Wi : j = 1, . . . , di ; j = 1, . . . , k}
is basis of W.
5
Proof. We only need to prove the last part. So, let
X
λij eij = 0
Then ωi ∈ Wi and
ω1 + ω 2 + · · · + ω k = 0 (I).
We will first prove that ωi = 0.
Since
T (eij ) = ci eij
for any polynomial f (X) ∈ F[X], we have
f (T )(eij ) = f (ci )eij .
Therefore,
di
X di
X
f (T )(ωi ) = λij f (T )(eij ) = λij f (ci )eij = f (ci )ωi (II).
j=1 j=i
Now let Qk
(X − ci )
g(X) = Qi=2
k
i=2 (c1 − ci )
Note g(X) is a polynomial. Also note this definition/ expression
makes sense because c1 , . . . , ck are distinct. And also g(c1 ) = 1 and
g(c2 ) = f1 (c3 ) = · · · = g(ck ) = 0.
Use (II) and apply to (I). We get
k
X k
X k
X
0 = g(T )( ωi ) = g(T )(ωi ) = g(ci )ωi . = ω1
i=1 i=1 i=1
6
Since Ei is a basis, λij = 0 for all i, j and the proof is complete.
7
2.13 (Definition) Let V be a vector space over F and V1 , V2 . . . , Vk
be subspaces of V. We say that V is direct sum of V1 , V2 , . . . , Vk ,
if each element x ∈ V can be written uniquely as
x = ω1 + ω2 + · · · + ω k
with ωi ∈ Vi .
Equivalently, if
1. V = V1 + V2 + · · · + Vk , and
V = V 1 ⊕ V2 ⊕ · · · ⊕ V k .
V = V 1 ⊕ V2 ⊕ · · · ⊕ V k
8
2.15 (Theorem) Suppose V is vector space of a field F with dim(V ) =
n. Let T ∈ L(V, V ) be linear operator. Suppose c1 , . . . , ck are the
distinct eigen values of T. Let
Wi = N (ci )
1. T is diagonalizable.
4. V = W1 ⊕ W2 ⊕ · · · ⊕ Wk .
9
3 Annihilating Polynomials
Suppose K is a commutative ring and M be a K−module. For
x ∈ M, we define annihiltor of x as
ann(x) = {λ ∈ K : λx = 0}.
10
4. Therefore,
ann(T ) = F[X]p(X)
where p(X) is the monic polynomial of least degree such that
p(T ) = 0.
This polynomial p(X) is defined to be the minimal monic
polynomial (MMP) for T.
f (A) = P f (B)P −1 .
p(c) = 0 ⇐⇒ g(c) = 0.
11
Recall g(X) = det(XI − A), where A is the matrix of T with
respect to some basis. Also by theorem 2.3, g(c) = 0 is and only if
cI − T is singular.
Now suppose p(c) = 0. So, p(X) = (X − c)q(x). for some q(X) ∈
F. Since degree(q) < degree(p), by minimality of p we have q(T ) 6= 0.
Let v ∈ V be such that v 6= 0 and e = q(T )(v) 6= 0. Since, p(T ) = 0,
we have (T − cI)q(T ) = 0. Hence 0 = (T − cI)q(T )(v) = (T − cI)(e).
So, (T − cI) is singular and hence g(c) = 0. This estblishes the proof
of this part.
Now assume that g(c) = 0. Therefore, T − cI is singular. So,
there is vector e ∈ V with e 6= 0 such that T (e) = ce. Applying this
equation to p we have
p | Q.
Proof. Write
Observe that
F ⊆ K ⊆ L(V, V ).
are subrings. Note Q(T ) ∈ K and we will prove Q(T ) = 0.
Let e1 , . . . , en be a basis of V and A = (aij ) be the matrix of T.
So, we have
12
(T (e1 ), T (e2 ), . . . , T (en )) = (e1 , e2 , . . . , en )A. (I)
Consider the following matrix, with entries in K :
T − a11 I −a12 I −a13 I · · · −a1n I
−a21 I T − a22 I −a23 I · · · −a2n I
B=
−a31 I −a32 I T − a33 I · · · −a3n I .
··· ··· ··· ··· ···
−an1 I −an2 I −an3 I · · · T − ann I
Therefore,
Therefore,
Q(T )(ei ) = 0 ∀i = 1, . . . , n.
13
4 Invatiant Subspaces
4.1 (Definition) Let V be a vector space over the field F and
T : V → V be a linear operator. A subspace W of V is said to be
invariant under T if
T (W ) ⊆ W.
4.2 (Examples) Let V be a vector space over the field F and
T :V →V
be a linear operator.
1. (Trivial Examples) Then V and {0} are invariant under T.
2. Suppose e be an eigen vector of T and W = Fe. Then W is
invariant under T.
3. Suppose λ be an eigen value of T and W = N (λ) be the eigen
space of λ. Then W is invariant under T.
4.3 (Remark) Let V be a vector space over the field F and
T :V →V
be a linear operator. Suppose W is an invariant subspace T. Then
the restriction map
T|W : W → W
is an well defined linear operator on W. So, the following diagram
T|W
W /W
² T
²
V /V
commutes.
4.4 (Remark) Let V be a vector space over the field F with dim(V ) =
n < ∞. Let
T :V →V
be a linear operator. Suppose W is an invariant subspace T and
T|W : W → W
is the restriciton of T.
14
1. Let p be the characteristic polynomial of T and q be the char-
acteristic polynomial of T|W . Then q | p.
and
Q(X) = det(Ir X − B).
Therefore Q | P. The proof is complete.
15
2. Now let V be a vector space over F with dim V = n < ∞. A
linear operator T : V → V is said to be triangulable, if there
is a basis E = {e1 , . . . , en } of V such that the matrix of V is
an (upper) triangular matrix. (Note that it does not make a
difference if we say ”upper” or ”lower” trinagular. To avoid
confusion, we will assume upper triangular.)
where ri ≤ di for i = 1, . . . , k.
16
(⇐): Now assume that the MMP p factors as
2. W1 is invariant under T.
3. dim V1 = n − 1 and V = W1 ⊕ V1 .
Let v ∈ V1 and T (v) = λ1 e1 + λ2 E2 + · · · + λn En , for some
λ1 , . . . , λn ∈ F. Define T1 (v) = λ2 E2 + · · · + λn En ∈ V1 . Then
T1 : V 1 → V 1
17
Therefore, the matrix of T m is given by
µ m ¶
m λ1 C m
A = .
0 Bm
2. W2 is invariant under T.
3. Also µ ¶
λ1 a12
(T (e1 ), T (e2 )) = (e1 , e2 ) .
0 λ2
1. dim(V2 ) = n − 2
2. V = W2 ⊕ V2 .
18
T2
V2 / V2
O
pr
²
T
V = W 2 ⊕ V2 / V = W 2 ⊕ V2
λ1 a12 a13
(T (e1 ), T (e2 ), T (e3 )) = (e1 , e2 , e3 ) 0 λ2 a23 .
0 0 λ3
Final Step: The process continues for n steps and we get linearly
independent set (basis) e1 , e2 , . . . , en such that
λ1 a12 a13 . . . a1n
0 λ2 a23 . . . a2n
(e1 , e2 , e3 , . . . , en )
0 0 λ3 . . . a3n .
... ... ... ... ...
0 0 0 ... λn
This completes the proof.
19
Recall a field F is said to be a an algebraically closed field if
every non-constant polynomial f ∈ F[X] has a root in F. It follows
that k is an algebraically closed field if and only if every non-constant
polynomial f ∈ F[X] product linear polynomials.
T : F n → Fn
p = (X − c1 )(X − c2 ) · · · (X − ck )
20
(⇐): We we asssume that p(X) = (X − c1 )(X − c2 ) · · · (X − ck ) and
prove thatPT is digonalizable. Let Wi = N (ci ) be a eigen space of ci .
Let W = ki=1 Wi be the sum of eigen spaces. Assume that W 6= V.
Now we will repeat some protions of the proof of theorem 4.6 and
get a contradiction. Let e1 , . . . , em be a basis of W and e1 , . . . , em , Em+1 , . . . , En
be a basis of V. Write V 0 = Span(Em+1 , . . . , En ). Note
1. W is invariant under T.
2. V = W ⊕ V 0 .
Define T 0 : V 0 → V 0 according to the diagram:
T0
V0 /V0
O
pr
² T
V =W ⊕V0 /V =W ⊕V0
21
5 Simultaneous Triangulation and Di-
agonilization
Suppose F ⊆ L(V, V ) is a family of linear operators on a vector space
V a field F. We say that F is a commuting family if T U = U T
for all U, T ∈ F.
In this section we try to find a basis E of V so that, for all T in
a family F the matrix of T is diagonal (or triangular) with respect
to E. Following are the main theorems.
Proof. We will omit the proof. You can work it out when you need.
22
6 Direct Sum
Part of this section we already touched. We gave the definition 2.13
of direct sum of subspaces. Following is an exercise. Note that we
can make the same definition for any subspace W.
6.2 (Examples) (1) R2 = Re1 ⊕ Re2 where e1 = (1, 0), e2 = (0, 1).
(2) Let V = Mnn (F). Let U be the subspace of all upper triangular
matrices. Let L be subspace of all strictly lower triangular matrices
(that means diagonal entries are zero). Then V = U ⊕ L.
(3) Recall theorem 2.15 that V is direct sum of eigen spaces of diag-
onizable operators T.
2. V = N ⊕ R.
Ei (v) = vi where v = v1 + · · · + vk , vi ∈ Wi .
23
Note Ei are well defined projections with
2. Ei Ej = 0 ∀ i 6= j.
3. E1 + E2 + · · · + Ek = I.
4. range(Ei ) = Wi .
Proof. The proof is easy and left as an exercise. First, try with
k = 2 operators, if you like.
24
7 Invariant Direct Sums
This section deals with some of the very natural concepts. Suppose
V is a vector space over a filed F and V = W1 ⊕ W2 ⊕ · · · ⊕ Wk .
where Wi are subspaces. Suppose for each i = 1, . . . , k we are given
linear operators Ti ∈ L(Wi , Wi ) on Wi , Then we can define a linear
operator T : V → such that
k
X k
X
T( vi ) = Ti (vi ) f or v i ∈ Wi .
i=1 i=1
² T
²
V /V
commute.
25
8 Primary Decomposition
We studied linear operators T on V under the assumption that the
characteristic polynomial q or the MMP p splits completely in to
linear factors. In this section we will not have this assumtion. Here
we will exploit the fact q, p have unique factorization.
1. V = W1 ⊕ · · · ⊕ Wk ;
Proof. Write
p Y r
fi = = pj j .
pri i j6=i
CGD(f1 , f2 , . . . , fk ) = 1.
Therefore
f 1 g1 + f 2 g2 + · · · + f k gk = 1
for some gi ∈ F[X].
For i = 1, . . . , k, let hi = fi gi and Ei = hi (T )) ∈ L(V, V ). Then
X
E1 + E 2 + · · · + E k = hi (T ) = Id. (I)
26
Also, for i 6= j note that p | hi hj . Since p(T ) = 0 we have
Ei Ej = hi (T )hj (T ) = 0. (II)
Ei T (V ) ⊆ Ei (V ) = Wi0 .
Therefore, Wi0 is invariant under T. We will show that Wi0 = Wi is
the null space of pi (T )ri .
We have
Since g(T )|W1 = g(T1 ) = 0, we have g(T ) vanishes on W1 and also for
j = 2, . . . , k we have pj (T )rj vanished on Wj . Therefore, g(T )f1 (T ) =
0. Hence p | gf1 . Hence pr1 = fp1 | g. Therefore pr1 is the MMP of T1
and the proof is complete.
27
Remarks. (1) Note that the projections Ei = hi (T ) in the above
theorem are polynomials in T.
(2) Also think what it means if some (or all) of the irreducible
factors pi = (X − λi ) are linear.
28