a first course in module theory
a first course in module theory
in Module Theory
This page is intentionally left blank
A First Course
in Module Theory
M E Keating
Imperial College, London
Distributed by
World Scientific Publishing Co. Pte. Ltd.
P O Box 128, Farrer Road, Singapore 912805
USA office: Suite IB, 1060 Main Street, River Edge, NJ 07661
UK office: 57 Shelton Street, Covent Garden, London WC2H 9HE
For photocopying of material in this volume, please pay a copying fee through the Copyright
Clearance Center, Inc., 222 Rosewood Drive, Danvers, MA 01923, USA. In this case permission to
photocopy is not required from the publisher.
vn
Vlll Introduction
and to lattices.
In Chapter 13, we use the module theory to find two standard forms for
a square matrix, namely, the rational canonical form and the Jordan normal
form of a matrix. In addition to the usual version of the Jordan normal
form of a matrix over the complex numbers, we give two further versions
that apply to a matrix whose entries are taken from a field other than the
complex numbers. The second of these variations is used, without proof,
in a fundamental paper on representation theory by J A Green [Green],
and a proof has not been published before at an elementary level, as far
as I know. I am grateful to my colleague Gordon James for drawing my
attention to these versions of the Jordan normal form.
In the final chapter, we go beyond the bounds of Euclidean domains to
look at some basic results on projective modules over rings in general. Here,
we establish some basic facts about group algebras and the relationship
between module theory and the representation theory of groups.
As befits a book that is intended as a first course for undergraduates,
arguments are given in considerable detail, at least, in the earlier part of
text. There is an index entry "proofs in full detail" to help the reader to
locate these agruments. There are many explicit illustrations and exercises,
and some hints and partial solutions to the exercises are provided in an
appendix.
Material that was not covered in the original lecture course is indicated
C by a "supplementary" symbol as in the margin. This material in not essen
tial for the core results on modules over Euclidean domains, and it can be
omitted if the reader wishes.
Each chapter has a section headed "Further developments" in which
the definitions and results of the chapter are placed in a wider setting.
References are provided to enable the interested reader to follow up the
topics that are introduced in these sections. I hope that these sections will
provide a useful source of projects for students who are in the final year of
a four-year undergraduate MMath or MSci course at a UK university, or
who are taking the new one-year preliminary research degree, the "MRes".
The text is divided into chapters and sections, which are numbered as
you would expect: 1.1, 1.2, and so on. For ease of reference, results are
numbered consecutively within each section, and so they appear in the form
"12.1.1 Theorem", "12.1.2 Lemma", etc.
This book was written while I was also working on the more advanced
texts [B & K: IRM] and [B & K: CM] with Jon Berrick of the National Uni
versity of Singapore. My collaboration with Jon has had many influences,
both on my teaching of the lecture course and on the composition of this
text, which it is impossible to acknowledge individually.
Introduction IX
Introduction vii
2 Euclidean D o m a i n s 17
2.1 The definition 17
2.2 The integers 18
2.3 Polynomial rings 18
2.4 The Gaussian integers 19
2.5 Units and ideals 20
2.6 Greatest common divisors 21
2.7 Euclid's algorithm 22
2.8 Factorization 23
2.9 Standard factorizations 25
2.10 Irreducible elements 26
2.11 Residue rings of Euclidean domains 28
2.12 Residue rings of polynomial rings rings 29
2.13 Splitting fields for polynomials 31
xi
xii Contents
4 Homomorphisms 51
4.1 The definition 51
4.2 Sums and products 53
4.3 Multiplication homomorphisms 54
4.4 F[X]-modules in general 55
4.5 F[X]-module homomorphisms 56
4.6 The matrix interpretation 57
4.7 Example: p = 1 57
4.8 Example: a triangular action 58
4.9 Kernel and image 58
4.10 Rank & nullity 60
4.11 Some calculations 61
4.12 Isomorphisms 62
4.13 A submodule correspondence 64
Exercises 65
5 Free Modules 69
5.1 The standard free modules 70
5.2 Free modules in general 71
5.3 A running example 73
5.4 Bases and isomorphisms 74
5.5 Uniqueness of rank 76
5.6 Change of basis 77
5.7 Coordinates 79
Contents xiii
9 Presentations 133
9.1 The definition 134
9.2 Relations 135
9.3 Defining a module by relations 136
xiv Contents
Bibliography 243
Index 245
Chapter 1
Each module has a ring of scalars associated with it. Therefore, we must
discuss rings before we can begin to tackle modules. In this chapter, we
collect together the basic definitions and properties of rings that we will
need in subsequent chapters. We consider ideals, which tell us about the
internal structure of a ring, and we look at some special types of ring,
particularly fields and polynomial rings. We also give the construction of
residue rings, which is an important method for obtaining new rings from
old.
A reader who has already met rings, ideals and Euclidean domains may
prefer to go directly to the start of our discussion of module theory in
Chapter 3, using this chapter and the next for reference.
We precede the definition of a ring with the definition of a more funda
mental structure, namely, a group.
1.1 Groups
We encounter two notations for groups, additive and multiplicative, which
are used according to the context in which the group arises. First, we
introduce the notation that is most often met in ring theory and module
theory.
An additive group is a (nonempty) set A together with a law of compo
sition + , called addition, which behaves as you might expect. Thus for any
a, b e A, there is an element a + b £ A which is called the sum of a and b,
and the following axioms must be satisfied.
1
2 Chapter 1. Rings and Ideals
Al: Associativity.
A2: Commutativity.
a + b = b + a for all a, b € A.
A3: Zero.
There is a zero element 0 in A with a + 0 = a for all a in A.
A4: Negatives.
Each element a oi A has a negative —a so that a + (—a) = 0.
It is usual to omit the bracket around a negative and write a + (—b) =
a — b and (—a) + b = —a + 6.
When studying groups in their own right, the law of composition is
usually written in multiplicative notation instead of additive notation. In
the multiplicative notation, a group is a (nonempty) set G in which any
elements g,h in G have a product g ■ h, often written simply as gh, and the
axioms are as follows.
Gl: Associativity.
G2: Identity.
There is an identity element 1 in G with g ■ 1 = g = 1 • g for all g in
G.
G3: Inverses.
Each element g of G has an inverse g~l so that
S - 9 " 1 = 1 = 0 - 1 •£■
Notice that our list of axioms for a multiplicative group is not simply
a translation of the list for an additive group. The difference is found in
axiom A2, which demands that an additive group must be abelian.
As groups are not our main concern in this text, we will not pause here
to give examples of them. We will meet additive groups in profusion in our
study of rings and modules, while multiplicative groups will be invoked to
construct some examples of rings in Chapter 14.
1.2. Rings 3
1.2 Rings
Informally, a ring is a set R in which arithmetic can be performed, that
is, the members of R can be added and multiplied, in much the same way
as integers or real numbers. However, there are two common properties of
multiplication that do not hold in a general ring.
The first of these properties is that a nonzero element of a ring need not
have a multiplicative inverse in that ring. For example, the integer 2 has
no inverse within the ring of integers Z, although it does have an inverse in
the rational numbers Q.
The second property is that multiplication need not be commutative,
that is, we can have rs ^ sr for two elements r and s of a ring.
Now for our formal definition of a ring. A ring is a nonempty set R,
on which there are two laws of composition, addition and multiplication.
Addition is indicated by the symbol +, so that for each pair r and s of
members of R there is a sum r + s in R. Multiplication is usually indicated
by simply writing the elements next to each other: for each pair r,s € R,
there is a product rs in R. When it is more convenient to have a symbol
for multiplication, we use a dot "•", so the product appears as r ■ s. We use
the dot while writing out the axioms.
Under addition, R must be an additive group as in the preceding section.
The properties of multiplication, and the interaction between addition and
multiplication, are given by the following axioms.
RM 1: Associativity.
RM 2: Identity.
There is an identity element 1 in R with
r ■ 1 = r = 1 • r for all r in R.
RM 3: Distributivity.
For all r, s and t in R,
r = rl = rO = 0 for every r 6 R,
so R must be the trivial or zero ring 0 that has only one element. Many
statements about rings or modules have trivial exceptional cases when they
4 Chapter 1. Rings and Ideals
are interpreted for the zero ring; as a rule, we will not state these exceptions
separately.
Familiar examples of rings are the ring of integers Z, the ring of rational
numbers Q, the ring of real numbers R and the ring of complex numbers
C. We shall assume the basic properties of these rings when we need to, as
any attempt to establish them in full detail would take us too far from the
point of this text. The more leisurely introduction to ring theory given by
Allenby [Allenby] does cover these topics.
s Given a ring R, the set Mn(R) of n x n matrices with entries in R is
also a ring under the usual addition and multiplication of matrices. The
verification of this assertion is a worthy but lengthy exercise.
1.4 Units
Let R be a ring. An element u of R is said to be a unit (or invertible) in R
if the following condition holds.
U: There is an element w in R so that
uw = 1 and wu = 1.
The inverse u _ 1 of a unit is itself a unit, with inverse u, and the product
of two units u, v is again a unit, with inverse v~lu~l. This means that the
set U(R) of units in R is a multiplicative group as in section 1.1.
Thus U(Z) = { + 1 , - 1 } , while U(Q) is the set of all nonzero rational
numbers. Notice that an element that is not a unit in one ring may become
a unit in a bigger ring.
An important type of ring is defined by the requirement that the zero
element is the only non-unit.
1.5 Fields
F. A field is a nonzero commutative ring F in which every nonzero
element is a unit of F.
The rational numbers Q, the real numbers R and the complex numbers
C are all examples of fields.
It is a fact that any commutative domain R is contained in a field Q
whose elements are of the form r ■ s~x = r/s for r , s € i J , s / 0 . The field
Q is called the field of fractions or quotient field of R. For instance, Q is
the field of fractions of Z.
As the existence of the field of fractions is extremely believable in all
the concrete examples that we consider, we will take it for granted. The
technical details of the construction are given in several texts - for instance,
section 3.10 of [Allenby] or [B & K: IRM] (1.1.12).
A discussion of the existence and construction of rings of fractions for
more general types of ring can be found in [Rowen] and [B & K: CM],
among other texts.
and
9 = go + giX + g2X2 + ■■■ + gnXn
in F[X], with m < n, their sum / + g is given by
where
1.6.1 Lemma
(i) If a ring F is commutative, so also is the polynomial ring F[X].
(ii) If a ring F is a domain, so also is the polynomial ring F[X].
Proof
Let
f(X) = fo + fiX + hX2 + ■■■ + fmXm
1.7. Ideals 7
and
9(X) =g0 + giX + g2X2 + ■■■+ gnXn
be in F[X], so that fo,fi,...,fm and go,gi,...,gn are in F; then their
product fg is the polynomial
h0 + hiX + --- + hkXk + ■■■ + hm+nXm+n
with
hk = fo9k H 1- fi9k-i H 1- Aito-
When F is commutative, the /c-th coefficient of / # is the same as the
fc-th coefficient of gf for k = 0,... ,m + n, since both are the sums of all
possible terms ftgk~i, i = 0 , . . . , k, but in different orders. Thus fg = gf
for all polynomials / , g, which gives the first assertion.
To prove the second, suppose that / and g are both nonzero. We may as
well assume that the polynomials have been written so that the coefficients
fm and gn are both nonzero. Then the product hm+n = fmgn is also
nonzero, so fg is nonzero. □
1.7 Ideals
Ideals are crucial to the investigation of modules. At an elementary level,
they provide the first examples of modules, and, at a much deeper level, a
knowledge of all the ideals of a ring sometimes enables us to describe the
modules over the ring. Ideals come in three types.
A left ideal of a ring R is a subset I of R with the following properties.
Idl: Zero. The zero element 0 of R is in I.
Id2: Additive closure. If x, y € / , then x + y e I.
IdL3: Multiplicative closure. If r £ R and x € I, then rx € / .
A right ideal I satisfies axioms Idl and Id2, but instead of IdL3 we have
IdR3: IfreR and x € I, then xr € / .
If I is simultaneously both a left and right ideal of R, we say that I is
a two-sided ideal.
If R is a commutative ring, then rx = xr for any x and r in R, so that
conditions IdL3 and IdR3 are the same. Thus every ideal of R is both left
and right and therefore two-sided. In this case, we refer simply to an ideal
of R. We will soon see an example in which left and right ideals differ from
one another.
For any ring R, R itself is a two-sided ideal of R. A (left, right or
two-sided) ideal / of R is called proper if I ^ R.
The subset {0} of R is a two-sided ideal, called, naturally enough, the
zero ideal of R. We usually write 0 for the zero ideal.
Next, we see how ideals of a ring arise from the elements of the ring.
8 Chapter 1. Rings and Ideals
ra + sa = (r + s)a € Ra.
By Lemma 1.8.1, these ideals are all distinct, and, in the next chapter, we
see that they are the only proper nonzero ideals of Z (Theorem 2.5.3).
Note: our usual rule is that we write a two-sided principal ideal as a left
ideal Ra. However, it looks unnatural to write Z2 in place of 2Z, etc.
g 4. For an example in which left and right ideals differ, we take R to
be the ring M 2 (R) of 2 x 2 matrices over the field R of real numbers. Let
{ 1 0
611 = J. Then
{ 0 0
Ren
= {{Z 2) | a i i ' a 2 i G R }
and
e R a
" = { { 7 o)\an,a12eRy
1.9. Sum and intersection 9
1.8.1 Lemma
Let R be a commutative domain, and let a and b be nonzero elements
of R. Then Ra = Rb if and only if a = ub where u is a unit of R.
In particular, Ra = R if and only if a is a unit of R.
I + J = {x + y | x € I, y e J},
7 n J = { x | a ; € / and x 6 J } .
1.9.1 Lemma
If I and J are both left ideals, right ideals or two-sided ideals of a ring
R, then I + J and I tl J are also correspondingly left, right or two-sided
ideals of R.
Proof
Suppose that I and J are both left ideals. First, we have 0 e I and
0 e J, giving 0 + 0 = 0 6 / + J .
Next, let x + y and x' + y' be members of / + J, with x,x' £ / and
y,y' e J. Then
(x + y) + ix' + y') = (x + x') + (y + y'),
10 Chapter 1. Rings and Ideals
1.10.1 L e m m a
Let I be a two-sided ideal of a ring R. Then congruence modulo I is an
equivalence relation on R.
1.10. Residue rings 11
r - 1 = (r - s) + (a -1) e I.
D
Given an element r of R, we define the residue class of r mod / to be
r = {s&R\s = r mod / } .
r + ~s = r + s
and
r • s = rs.
The hardest point in the verification that these laws of composition make
R/I into a ring is to check that they are well-defined. This problem arises
since one element of R/I can be expressed as the residue class of many
different elements from R, and we need to know that the sum and product
in R/I are not affected by such variations. We record the statement and
proof as follows.
1.10.2 Proposition
Let I be a two-sided ideal of a ring R. Then
(i) addition and multiplication in R/I are well-defined;
(ii) R/I is a ring with zero element 0 and identity element 1;
(Hi) if R is commutative, then R/I is also a commutative ring.
Remark: Exercise 1.8 gives a nontrivial example in which R/I is commu
tative although R is not.
12 Chapter 1. Rings and Ideals
Proof
(i) Suppose that r = ?i and s = si, so that r = n + x and s = si + y for
elements x, y € /.
Then r + s = (ri+si) + (x + y) with x + y € I and hence r -f s = T\ + Si,
that is, the two alternative methods for computing the sum F + s give the
same result.
Also, rs = r i s i + (r x y + s\x + xy) with rxy + six + xy € I, so the two
computations of r • s have the same outcome.
(ii) This is a matter of checking that the addition and multiplication in R/I
satisfy the axioms as given in section 1.2, granted that the axioms hold in
R already. We give two sample checks to show how easy it is.
For any r e f l ,
r+ 0 = r +0
= f
(r + s) -t = r + s -t
= (r + s)t
= rt + st
= ri + li
= r -t + s -t
which establishes that the distributive law holds.
(iii) Suppose R is commutative. Then, for any elements f,s of R/I, we
have
r-~s=f~s = ~sr = ~s-r.
D
To describe Z m , we use the fact that, for any integer s, there are integers
q and r with s = qm + r and 0 < r < m — 1.
Now, for integers r, s,
Z m = { 0 , T , . . . , m - 1},
and these are all distinct since m cannot divide r — s if 0 < r, s < m — 1
unless r = s.
Note that the ring Z m need not be a domain even though Z is a domain
- for example 2 • 2 = 0 in Z4, although 2 ^ 0 . However, in the important
case when p is a prime number, the ring Z p is a field, which we deduce from
a more general result that is our next theorem.
Before we can state this theorem, we need a preliminary definition. A
two-sided ideal I of a ring R is maximal if I is a proper ideal of R, and
there is no two-sided ideal J of R with I C J C R. Here, we use the symbol
"C" to indicate strict containment, that is, I ^ J and J / R. Maximal
left ideals and maximal right ideals are defined in the obvious way.
Granted the unique factorization of integers, and that every ideal in Z
is principal, the maximal ideals in Z have the form pZ where p is prime.
1.11.1 T h e o r e m
Let I be an ideal of a commutative ring R. Then the following state
ments are equivalent.
(i) I is a maximal ideal of R.
(ii) The residue ring R/I is a field.
Proof
(i) => (ii): By Lemma 1.10.2, R/I is a commutative ring, so we only need
to find an inverse for a typical nonzero element f € R/I. By Lemma 1.9.1,
Rr+I is an ideal of R, and, as / is maximal, either Rr+I = I or Rr+I = R.
But if Rr + I = I, then r = l-r + 0el and so f = 0, contradicting our
assumption that r is nonzero.
Thus Rr + I = R, and so there is an element s of R and an element x
of I with sr + x = 1. This equality gives s • f = 1 in R/I, that is, f has an
inverse.
(ii) => (i): Suppose that J is an ideal of R with I C J, (so I ^ J). Then
there is some element r £ J with r $ I. Since F ^ 0 in R/I, s • f = 1 for
some s € R.
14 Chapter 1. Rings and Ideals
1.11.2 Corollary
Let p be a prime number. Then Z p is a field. D
Exercises
1.1 A nonzero element r of a ring R is called a proper zero divisor if
rs = 0 for some nonzero element s £ R. Show that a proper zero
divisor cannot be a unit of R.
1.2 Show that a commutative ring R is a field if and only if 0 is the only
proper ideal of R.
1.3 Let R be a commutative ring. An ideal / of R is said to be prime if
the following holds.
P: If r, s £ R and rs £ I, then either r £ I or s £ I (possibly both are
in / ) .
Show that
(i) 0 is a prime ideal of R «=>• R is a domain.
(ii) / is prime <=> R/I is a domain.
(iii) A maximal ideal of R must be a prime ideal.
1.4 Let m > 2 be an integer. Show that Z m is a domain <=>• m is a
prime number.
Show by direct calculation that U(ZQ) = {1,5} and that the set
of proper zero divisors in I,Q is {2,3,4}.
Find the corresponding results for Zg and Zio-
1.5 Let A be a commutative domain and let R = A[X] be the polynomial
ring over A. Prove the following assertions.
(a) / = f{X) = /o + fiX + ■ ■ • + fnXn is a unit of R <=> n = 0
and /o is a unit of A.
(b) X divides a product fg in i? <=>■ X divides either / or g.
(c) X = fg with f,g £ R «=*> * either / or g is a unit of R.
1.6 Let F be a field and let fl = F[X, Y] (= F[X][Y]) be the polynomial
ring in two variables over F. Show the following.
(a) R is a domain.
(b) RX is a prime ideal of R.
(c) i?X + RY is a maximal ideal of R but .RX + RY is not principal.
g 1.7 Let F be a field and let D be the set of all diagonal matrices ( ^ ]
with r,s € F. Verify that D is a commutative ring under the usual
sum and product of matrices.
Exercises 15
Show also that the only proper nonzero ideals of D are De and
Df, where e =( J jj J and / = ( J J J, and that De + D / = D
and DedDf = 0.
Show that there is a bijective map 9 : D/De —► F given by
r 0
x 0 s
and that
for all d, d' in D. Thus we can identify the ring D/De with F.
Remark in other words, 9 is an isomorphism of rings - see Exercise
4.4.
1.8 Let F be a field and let T be the set of all upper triangular matrices C
/ r t \
I n . ) with r, s,t in F. Verify that T is a ring under the usual
sum and product of matrices. Show that T is not commutative.
Let
H
={{I $) | r , t 6 i r }
0 t
t,seF
0 s
'-{(SO1"'}
Show that i7, J and J are all two-sided ideals in T. Using the
methods of the preceding exercise, show that T/H = T/I = F, while
T/J = D. (Thus, as promised earlier, a noncommutative ring can
have a commutative residue ring.)
1.9 Let F be a field and let R be the ring of all 2 x 2 matrices over F .
Show that R has no two-sided ideals except 0 and R. (See Exercise
7.8.)
Chapter 2
Euclidean Domains
17
18 Chapter 2. Euclidean Domains
The element q is called the quotient and r the remainder. Either of them
may be 0, and they may not be uniquely determined by the pair a, b.
Axiom ED 3 is called the division algorithm, or, less formally, long
division.
Any field F is a trivial Euclidean domain in which <p(x) = 1 for any
nonzero element x of F. We next consider some nontrivial examples.
Z m = {0, . . . , m - 1}.
Note also that axiom ED 3 does not define the residue uniquely - we
are permitted to write 7 = 3 - 2 + 1 o r 7 = 4 - 2 + (—1) when we divide 7
by 2, for instance. However, custom dictates that we choose the smallest
non-negative residue.
nJ>
1 o, / =o
2.4. The Gaussian integers 19
and
(a + bi) ■ (c + di) = (ac — bd) + {ad + bc)i
respectively, are again Gaussian integers.
The fact that Z[i] is a ring can now be confirmed, in one of two ways,
according to the reader's inclination. The hard but secure route is to check
all the axioms, one by one. A simpler method, which tends to arouse
suspicion at first sight, is to observe that the set C of complex numbers is a
ring (which we take for granted), and that properties such as associativity,
distributivity,..., are therefore inherited by Z[»]. Similarly, there are two
ways to verify that Z[i] is a commutative domain.
For each Gaussian integer a + bi, define
2.5.1 Lemma
Let R be a Euclidean domain and let u be in R. Then u is a unit of R
if and only if f(u) = 1.
<p(u)<p(v) = 1 in Z.
2.5.2 Corollary
(i) In the Gaussian integers Z[i], the units are { ± l , ± i } .
(ii) In a polynomial ring F[X], F a field, the unit polynomials are the
nonzero constant polynomials / = /o- □
Next we establish the key result about the ideals of a Euclidean domain.
Recall that an ideal I of R is principal if / = Ra = {ra \ r G R} for some
element a of R.
2.6. Greatest common divisors 21
2.5.3 Theorem
Let R be a Euclidean domain. Then every ideal of R is principal.
Proof It is obvious that the zero ideal 0 is principal, its unique generator
being the zero element 0. Now suppose that 7 ^ 0 , and choose an element
a of I so that
0 < <p(a) < ip(x) for all x £ I.
Clearly, Ra C I. To prove equality, take any x £ I and write x = qa + r
with <p(r) < ip(a). Since r = x — qa, r belongs to I. But the inequality for
<p(a) now forces the equality <p(r) = 0 and so r = 0. □
GCD 1: d | a and d | b,
GCD 2: d = sa + tb for some elements s,t of R.
2.6.1 L e m m a
Let R be a Euclidean domain and let a,b G R. Then d € R is a greatest
common divisor of a and b <*=> Rd = Ra + Rb.
In particular, a, b always have a greatest common divisor, which can be
written d = sa + tb for some s,t € R.
Then
rk - (rk-i,rk) = •■• = (a,b).
2.8 Factorization
Our aim now is to show that, in an arbitrary Euclidean domain, there is a
unique factorization theorem analogous to that for the integers. We must
first extend some definitions from Z to Euclidean domains.
An element p of a commutative domain R is said to be irreducible if p
is neither zero nor a unit, and whenever p = ab for a, b in R, then either
a or b is a unit. When we are working in Z, we defer to tradition and say
'prime number' in preference to 'irreducible number'.
A nonzero, nonunit element which is not irreducible is called reducible.
If p is irreducible, so also is up for any unit u of R. Two elements x, y of
R are said to be associates if x = uy for some unit u of R, and such elements
are not regarded as being genuinely distinct when computing factorizations.
Thus, when the irreducible elements p, q are said to be distinct irreducible
elements, it is to be understood that they are not associates, rather than
simply being unequal.
In Z, prime numbers appear in pairs of associates {2, —2}, {3, — 3 } , . . . ,
and custom dictates that we always choose the positive prime. In a poly
nomial ring F[X] over a field F, irreducibles occur in families of associates
{up | u € F, t i / 0 } , since the unit polynomials are the nonzero constants.
Here, the custom is to choose the unique monic polynomial in each family
of associates.
Two elements a, b of R are said to be coprime if their greatest common
divisor (a, b) is 1 (or a unit, which can always be replaced by 1). The
following sequence of results is the key to unique factorization, and also to
some other important results that we will meet later.
2.8.1 P r o p o s i t i o n
Let R be a Euclidean domain and suppose that a, b G R. Then a,b are
coprime •$=> 1 = sa + tb for some s,t € R.
2.8.2 Proposition
Let R be a Euclidean domain and suppose that a,b S R are coprime. If
a | be with c £ R, then a | c.
2.8.3 Corollary
Let R be a Euclidean domain and suppose that a,b € R are coprime.
Then RaC\Rb = Rab. □
2.8.4 Corollary
Suppose that p G R is irreducible and that p \ be where b and c are in
R. Then either p\b or p\c.
{P2,--,Ps}
and
{qi,.. .,<7j_i,<7j+i,... ,qt}
as required. Q
where u is a unit and pi, ■ ■ ■ ,Pk are distinct (that is, non-associated) irre-
ducibles.
Such a factorization is called a standard factorization of a. The unique
ness part of the theorem above tells us that the set of irreducibles pi,...,pk
is uniquely determined by a (apart from the order in which it is written),
and that the exponents n ( l ) , . . . ,n(fc) are uniquely determined by a once
an order of listing for irreducibles is fixed.
26 Chapter 2. Euclidean Domains
P r i m e numbers
We are obliged to assume that we know a prime number when we see
it. Although we know that there are infinitely many prime numbers (see
Exercise 2.1), we cannot list any infinite subset of them, and the problem
of determining very large primes is an active area of research.
f = (X-X1)---(X-Xn).
/ = (X - Xx)(X - A2).
Writing b = —(Ai + A2) and c = A1A2 gives the required form for / .
2.11.1 Proposition
Let R be a Euclidean domain, I an ideal of R. Then the residue ring
R/I is a field <?$■ I = Rp with p irreducible.
Proof
By Theorem 1.11.1, R/I is a field if and only if I is a maximal ideal of
R. But I = Ra for some element a of R, and it is clear that Ra is maximal
precisely when a is irreducible. □
2.12. Residue rings of polynomial rings 29
90 + 9ie + 92£2 H 1- 3 n - i « n _ 1
where e is a root of / . The addition and multiplication of such polynomials
follows the expected rules, with the relation /(e) = 0 being used to eliminate
the powers en,en+1, ... from products.
Before we get into the technicalities, here is a familiar example. Take F
to be the field R of real numbers, and let / = X2 + 1. Then the elements
of RfXj/Rpf]/ have the form go + gie, where e2 = —1. Apart from nota
tion, this is the classical definition of the complex numbers as pairs of real
numbers.
Now we start our formal analysis of the residue ring F[X]/F[X]f. First
we note that we have not lost any generality by imposing the requirement
that / is monic, since the ideal F[X]f is unchanged if we replace / by uf
for any nonzero constant u (see Lemma 1.8.1 and Corollary 2.5.2).
There is no loss either in requiring that deg(/) > 1, since when / is
a nonzero constant polynomial, F[X]f = F[X] and so F[X]/F[X]f is the
trivial ring.
Now let g € F[X) be arbitrary. As we saw in section 2.3, we can write
g = qf + r
with the remainder r having degree less than n. Since g = f in F[X\/F[X]f,
the elements of F[X]/F[X]f can be taken to have the form g with g £ F[X],
deg(g) <n.
If his also a polynomial with deg(/i) < n and g = h, then g — h 6 F[X]f,
so / divides g — h. But deg(g — h) < n, so the division is possible only
if g - h = 0. It follows that each element of F[X]/F[X]f can be written
uniquely in the form ~g with deg g < n.
Suppose then that degg < n and write
g = g0 + giX + g2X2 + ■ ■ ■ + gn-xX71-1, g0,...,gn-i € F,
where any (or indeed all) of the coefficients gi can be 0 - the notation is
not to be interpreted to imply that deg(<?) is n— 1. Then, in F[X]/F[X]f,
g = Ifi + giX + g2X2 + hffn-iX™"1
30 Chapter 2. Euclidean Domains
m = / = o,
so that the multiplication in F[X]/F[X]f follows from polynomial multi
plication, together with the rule
where
f = fo + fiX + --- + / n - i X " - 1 +Xn,n = deg(/).
Note: the "root" e that we have constructed for the polynomial / is an
algebraic entity. Even if / has real coefficients, it may not be possible to
identify e as a real or complex number. For example, if / = X2, then
e2 = 0 but e -fi 0. This observation illustrates the fact that the residue ring
F[X]/F[X]f will not be a field if / is not irreducible.
2.13. Splitting fields for polynomials 31
2.13.1 Proposition
Let F be a field and let f be a polynomial with coefficients in F. Then
there is a field E containing F in which f is split-
Proof Write the irreducible factorization of f(X) in F[X] as
f(X) = (X-\1)-.-(X~ Afc)pi(X) ■ • -p,(X)
where we have gathered all the linear factors of f(X) at the start. We allow
the possibilities k = 0 or s = 0, and the Aj need not be distinct.
Let n = deg(/) and put d = d(f,F) = n — k. We induce on d. The
initial case is that d = 0. Then k = n, so we must have s = 0, that is, F is
already a splitting field for / .
Suppose that d > 0, so that s > 1. Let F' = F[X]/F[X]pi be the ring
constructed in section 2.12. Then F' contains F. Further, by Proposition
2.11.1, F' is a field, and Equation 2.2 shows that F' contains a root of p\.
Thus the factorization of / over F' has more than k linear terms, so that
d(f,F')<d(f,F).
By our induction hypothesis, there is a field E containing F' in which
/ splits. □
Remarks If F is contained in the field C of complex numbers, the Fundamen
tal Theorem of Algebra shows that C is a splitting field for any polynomial
over F. However, the algebraic construction for the splitting field works for
any field F.
A stronger result can be proved, that a polynomial has an essentially
unique smallest splitting field ([Allenby], p. 149), but we do not need this
for our purposes.
This weaker axiom has the advantage that it allows more rings to be
considered to be Euclidean domains than does ED 2, while causing only
a small increase in the complexity of proofs. However, all the Euclidean
domains that we meet in this text satisfy the stronger axiom.
• The technique used to show that the Gaussian integers are a Euclidean
domain can be extended to some similar types of ring - see Exercise 2.8
below and [Allenby] §3.7 for some of the easier cases. Comprehensive dis
cussion of the limits of the technique are given in Chapter 14 of [H & W]
and in [E, L & S].
• A commutative domain in which every ideal is principal is called a prin
cipal ideal domain. By Theorem 2.5.3, a Euclidean domain is a principal
ideal domain. The Unique Factorization Theorem holds over a principal
ideal domain, as do most of the results that we subsequently obtain for
modules over Euclidean domains, but the arguments are more technical.
As genuine (that is, non-Euclidean) principal ideal domains are rarely en
countered in undergraduate mathematics, and not too often at any level,
this text keeps to Euclidean domains. The proof of unique factorization
in a principal ideal domain can be found in section 2.15 of [Jacobson] or
section 10.5 of [Cohn 1], among others.
The most accessible non-Euclidean principal ideal domain is the ring
Z[(l + \/—19)/2]. The fact that this ring is not Euclidean is proved in
the sources mentioned in the previous paragraph, while some algebraic
number theory is needed to show that it is nevertheless a principal ideal
domain. [Marcus] gives a nice account of the calculations needed; Exer
cise 9 of Chapter 5 is particularly relevant.
• The results of this chapter (and book) can be extended to noncommu-
tative versions of Euclidean domains ([B k. K: IRM], Chapter 3) and
to noncommutative principal ideal domains ([Cohn: FRTR], Chapter 8;
[Rowen]).
Exercises
2.1 Suppose that pi,- ■ ■ ,Pk are distinct prime numbers. Show that the
product
Pi•■•Pk + 1
has a prime factor q with q ^ Pi for any i. Deduce that there are
infinitely many prime numbers.
2.2 Let f(X) and X - A be polynomials over a field F. Show that f(X) =
q{X)(X - A) + r where r G F. Deduce that A is a root of f(X) if
and only if X - A | f{X). Prove further that f(X) can have at most
deg(/) distinct roots in F
Exercises 33
2.3 Let p e Z b e prime. Arguing directly from Gauss's Lemma, show that
the polynomials Xn — p are all irreducible for n > 2.
Generalize your argument to a proof of Eisenstein's Criterion.
2.4 Let R be a Euclidean domain and let a in R be neither a unit nor
irreducible. Show that the ring R/Ra contains a nontrivial divisor of
0.
2.5 Let f(Y) = f(Y + 1) be the polynomial obtained by the change of
variable X = Y +1 from f(X) € F[X], F a field. Show that fg = fg.
Deduce that f{X) is irreducible if and only if f(Y) is irreducible.
Let p € E be a prime number. Prove that the polynomial Xp~ * +
• • ■ + X + 1 = (X" - 1)/(X - 1) is irreducible in Q[X].
2.6 Show that the polynomial ring F[X, Y] in two variables over a field is
not a Euclidean domain.
2.7 Let R = Z[%/—5]. Show that 3, 2+\J—5 and 2 — yj— 5 are all irreducible
elements of R and that no two of them are associates. Verify that
3 2 = (2 + v / ^ 5 ) ( 2 - V ^ 5 ) .
This means that unique factorization does not hold in R and hence
that R cannot be a Euclidean domain. Confirm this by showing that
the ideal 3R + (2 + y/—5)R is not principal.
2.8 Some more Euclidean domains. Here are some rings which can be
shown to be Euclidean domains using mild variations of the technique
for the Gaussian integers.
i: Z[yf^\, <p(a + by/^2) = a2 + 2b2.
ii: Z[w], where u> = (— 1 + -</—3)/2 is a cube root of 1, and
Now that we have finished our introductory survey of rings, we can intro
duce the main objects of our enquiries in this text, namely, modules. This
chapter is concerned with the definitions and general properties of mod
ules and their submodules, together with concrete examples arising from
additive groups and from matrices acting on vector spaces. The latter type
of module will prove to be crucial to our investigation of normal forms for
matrices in Chapter 13.
Although we are mainly interested in modules over commutative do
mains, we allow the ring of scalars to be arbitrary in our basic definitions.
35
36 Chapter 3. Modules and Submodules
\vn )
38 Chapter 3. Modules and Submodules
/ n \ ( Wi \ ( Vi + VJi >
V2 + W2
V2 W2
+
\Vn ) \ wn / \ vn + wn )
and
/ vi \ ( kVl \
V2 kv2
V vn j \ fc
^n /
where k is in F.
Now suppose we have a n n x n matrix
/ a n c-12 ain \
A= O-il 0-i2
which again belongs to Fn; this is what we mean when we say that "the
matrix A acts on the space Fn ".
Careful calculation shows that for any such matrix A, any vectors v, w
in Fn and any scalar k in F, we have
A(v + w) = Av + Aw
and
A(kv) = k(Av),
that is, A acts as an F-linear transformation on Fn.
3.4. Actions of scalar matrices 39
Since the powers A2, A3,..., A1,... of A are also nxn matrices over F,
they all act on F , giving vectors A2v, A3v,..., Aiv,... for v e Fn.
n
put
fv = f0v + fxAv + ■■■ + fiA{v + ■■■ + fnAnv for v e Fn.
With the convention that ^4° = / , the nxn identity matrix, we can write
fv as the sum
fv = YdfiAiv.
i=0
A great deal of checking (which is left to the reader as a very good exercise)
confirms that Fn becomes an F[X]-module M with this scalar multiplica-
tion.
We use the expression "M is given by X acting as A" to indicate that
the F[X]-module M is Fn made into an F[X]-module by this construction.
The value of n, the choice of F and the fact that A is an n x n matrix over
F will usually be clear from the context.
Note that each choice of an n x n matrix gives a different module struc-
ture on Fn. For this reason, we sometimes use the notation M(A), M(B),...
to indicate the modules given by X acting as the matrices A, B,....
Here are some uncomplicated examples.
is
fv = f0v + fiXv + ■■■ + fiX{v + ■■■ + fnXnv, for v e Fn.
An important special case occurs when A = 0, that is, A is the zero nxn
matrix. Then Xv — 0 always, and fv = f0v for any polynomial / . In this
40 Chapter 3. Modules and Submodules
3.5 Submodules
Before we discuss some further explicit examples of modules, we take a first
look at the internal structure of a general module.
Let M be a left module over an arbitrary ring R. An R-submodule of
M is a subset L of M which satisfies the following requirements.
SubM 1: 0 6 L.
SubM 2: If I, /' € L, then I + I' e L also.
SubM 3: HI € L and r £ R, then rl € L also.
A submodule L of a left .R-module is itself a left .R-module, since the
axioms for addition and scalar multiplication already hold in the larger
module M.
If the ring R of scalars can be taken for granted, we omit it from the ter
minology and say that L is a submodule of M rather than an i?-submodule.
An .R-submodule of a right module M is defined by making the obvious
modification to axiom SubM 3:
SubMR 1: HI € L and r £ R, then Ir € L also.
Clearly, a submodule of a right /^-module is again a right module.
General statements and definitions about submodules of left modules
have obvious counterparts for submodules of right modules. As our main
interest is with modules over commutative rings, for which we use the left-
handed notation, we will as a rule give only the left-handed versions of
such statements and definitions. The reader should have no problem in
providing the right-handed versions where desired.
3.6. Sum and intersection 41
When the ring of scalars R is commutative, then any left module M can
be regarded as a right module, and any submodule of M as a left module
is equally a submodule of M as a right module.
A left module M is always a submodule of itself. A submodule L of M
with L / M is called a proper submodule of M.
At the other extreme, any module has a zero submodule {0}, which we
usually write simply as 0. The zero submodule is a proper submodule of
M unless M is itself the zero module.
A left module S is said to be a simple module if S is nonzero and it has
no submodules except 0 and itself. The description of the simple modules
over a ring R is one of the fundamental tasks in ring theory.
Next, we look at some situations where submodules are already known
to us under different names.
3.6.1 Lemma
(i) Both L + N and LC\ N are submodules of M.
(ii) L + N = L <i=^> N C L.
(Hi) LnN = L <=> LCN.
m + m' £ L + N
since
l + l' € L and n + n' € N.
3.8 Generators
A convenient way to specify a module or one of its submodules is in terms
of generators. In fact, our investigation of the structure of modules over
Euclidean domains will be based on an analysis of the generators of the
modules. Again, we work only with left modules, and leave right-handed
definitions to the reader.
First, we consider a single generator. Let M be a left .R-module and
let x be an element of M. The cyclic submodule of M generated by x is
defined as
Rx = {rx | r e R}.
The left module M itself is said to be cyclic if M = Rx for some x e M,
and the element x is called a generator of M.
The confirmation that Rx is actually a submodule of M is a trivial
extension of the argument given for principal ideals in section 1.8; also, in
a moment we will give a more general calculation which includes the cyclic
case.
Notice that a principal left ideal Rx is a special type of cyclic submodule,
where we take M = R and x in R.
We next consider finite sets of generators. Let X = {x\,... ,xt} be a
finite subset of M, and put
L{X) = { n i l H \-rtxt | ri,...,rt 6 R},
r ■ (rii! H h rtxt) = (r • r i ) X\ + h (r • r t ) i t
For example, the sets X = {1} and Y = {2,3} are both generating sets
of Z, considered as a module over itself. The set X is linearly independent
in an obvious sense, but Y is not. Further, neither element of Y can be
omitted to give a generating set with one member.
3.9.1 T h e o r e m
Let F be a field, let A be an n x n matrix over F, and let M be the
F[X)-module obtained from the vector space Fn with X acting as A. Then
there is a bijective correspondence between
(i) F[X)-submodules L of M,
and
46 Chapter 3. Modules and Submodules
3.10 Eigenspaces
Given an n x n matrix A over a field F, an eigenspace for A is a nonzero
subspace U of Fn with the property that there is a scalar A g F s o that
Au = Aw for all u € U.
3.10.1 Lemma
Let F be a field, let A be annxn matrix over F, and let M be the F[X]-
module obtained from the vector space Fn with X acting as A. Further,
suppose that the subspace U of Fn is one-dimensional over F.
3.11. Example: a triangular matrix action 47
"'-- \ l I ^ A/>
x+y
Xm-i
y
A proper subspace of F2 must have dimension 1, and hence a proper sub-
module L of M must be given by an eigenvector of A. The eigenvalues of
A are the roots of
(o
for x / 0.
It follows that there is exactly one proper submodule of M, given by
the subspace U = F ( _ j of F2.
Noice that this result is not influenced by any specific properties of the
field F - it holds if we take F to be K, C, Z p where p is a prime, or any
other field.
48 Chapter 3. Modules and Submodules
Xm =
v+ = ( I and V- = (
l
\ J V -»
We obtain two one-dimensional C-submodules of M, namely L + = Cv+
and L_ = Cv_.
Exercises
3.1 Show that the set {6,10,15} generates Z as a Z-module, but that no
subset of {6,10,15} generates Z.
For each k > 4, find a set A). = {ai,.,.,afe} of integers which
generates Z, so that no subset of Ak generates Z.
3.2 Let p e Z be a prime number. Show that if £ is a nonzero element
of the additive group Z p , then x is a generator of Z p as a Z-module.
Deduce that Z p is a simple Z-module. Hint: Corollary 1.11.2.
Remark this result is more frequently given in the form "a cyclic
group of prime order has no nontrivial subgroups".
Exercises 49
rA = {ra | a G >!}
is a subgroup of A.
Take A - Z m - Prove that rA = A <=> (r, m) = 1 and that
rA = 0 <=> m divides r.
More generally, let R be a Euclidean domain and let M be an
.R-module. Given an element r e R, show that
rM = {rm | m £ M }
is a submodule of M.
Suppose that M = R/Rx for some x e R. Determine conditions
on r so that (i) rM = M and (ii) rM = 0.
3.4 Repeat Example 3.12 for the finite fields Z p , p prime. (Note that
y/—l € Z p <=> p = 1 mod 4 for odd p.)
3.5 Let M be the C[X]-module given by an n x n matrix A acting on C n .
Show that M is simple if and only if n = 1.
/0 1 1\
3.6 Let M be the C[X]-module given by the matrix A = I 0 0 1 I
\ 0 0 0 /
acting on C 3 . For each vector v € C 3 , let L(v) be the cyclic C[X]-
submodule of M generated by v, and write L0 for L{e\) where e\ =
1
0
0
Show that Cei is the only eigenspace of C 3 . Deduce that LQ C
L(v) for any v ^ 0. Find all v with
(a) dimL(v) = 2,
(b) dimL(v) = 3.
Do your answers change if the field C is replaced by an arbitrary
field F?
/ 0 1 0
3.7 Let N be the C[X]-module given by the matrix B = j 0 0 1
\ 1 0 0
acting on C 3 .
(a) Prove that there are exactly 3 submodules of iV that are one-
dimensional as vector spaces over C.
(b) Show that TV is a cyclic C-module.
n
of the form ( 021 „0 |. Show that, under the usual rules for matrix
multiplication, / is a right ideal of R (and hence a right R-module)
but that I is not a left ideal. Show also J is a left ideal but not a
right ideal in R.
Prove that I is simple as a right module and that J is simple as a
left module.
Generalize these results to the ring o f n x n matrices over F. (See
also Exercise 7.8.)
Chapter 4
Homomorphisms
51
52 Chapter 4. Homomorphisms
We will work only with left modules and their homomorphisms in this
chapter.
Alternative terms are module homomorphism, when there is no doubt
about the choice of coefficient ring R, or simply homomorphism if it is
obvious that we are dealing with modules rather than groups, rings, or
some other mathematical structure.
When the ring of scalars is a field F, so that M and N are then vector
spaces over F, an F-module homomorphism is more familiarly known as
an F-linear transformation or an F-linear map, or simply a linear transfor
mation or linear map.
Some authors extend the vector space terminology to modules in gen
eral, and speak of ".R-linear transformations" or ".R-linear maps". However,
as we shall be concerned with the relationship between F-linear maps and
F[X]-module homomorphisms when we analyse the structure of modules
over polynomial rings, it will be convenient to limit the use of the term
"linear" to vector spaces.
When M and N are Z-modules, that is, additive groups, the second
axiom HOM 2 follows automatically from the first. Thus a Z-module ho
momorphism is another name for a group homomorphism from M to N.
(The reader who has not studied group theory can take this as a definition.)
Here are three homomorphisms that are always present.
idM : M ->■ M
is defined by
idM (TTI) = m for all m € M.
inc : L ->■ M,
defined by
inc(l) = I for all / e L.
At first sight, it may seem pointless to give names to these "do nothing"
maps, but there are circumstances where it is very useful to be able to
distinguish between an element of L regarded as an element of L and the
same element regarded as an element of M.
4.2. Sums and products 53
and
(9 + ip)(rm) = 9{rm) + ip(rm)
= r ■ 9(m) + r ■ ip(m)
= r ■((<? +V0(m))-
M = {0,1,2,3,4,5},
we have
a(a)M = {0,a, 2a, 3a, 4a, 5o}.
4.4. F[X]-modules in general 55
a:V -> V
a* : V -> V
is an F-linear transformation for any i > 1, and we can define a scalar
multiplication of polynomials
on vectors v in V by
A great deal of checking, which is left to the reader, confirms that V has
become an F[X]-module. We give this a new name, M, and say that M is
defined by X acting as a on V.
56 Chapter 4. Homomorphisms
9a = 130.
and so
6{Xi ■m)=Xi ■ 6(m) for all m and all i > 1.
Thus, for any polynomial f(X) = f0 + fiX + h fsX3 and any m € M,
4.5.1 T h e o r e m
Let F be a field, and suppose that the F[X]-module M is given by the
action of the F-linear transformation a on the F-space V and that N is
given by the action of P on W.
Then there is a bijective correspondence between
(i) F[X]-module homomorphisms 9 : M —> N
and
(ii) F-linear transformations 6 : V —»■ W such that 6a = /39. □
4.7 Example: p = 1
As an illustration, we look at the elementary (but sometimes confusing)
case in which p = 1 but n is arbitrary. The space F = F1 has dimension 1,
and we can take the single element 1 G F to be a basis. The action of the
58 Chapter 4. Homomorphisms
0 in
TA
0 i2i
and
*21 *22
AT =
0 0
so T gives a homomorphism precisely when i n = i22 and i2i = 0, that is,
T = I . I with i n and ii2 arbitrary.
Ker(0) = {m e M | 9{m) = 0}
4.9. Kernel and image 59
The corresponding definitions for right modules are left to the reader.
4.9.1 Lemma
(i) Ker(0) is a submodule of M.
(ii) 0 is injective •<=> Ker(0) = 0.
(Hi) \m(6) is a submodule of N.
(iv) 6 is surjective <=> Im(6) = N.
Remark. Some texts use the terms one-to-one for an injective map and
onto for a surjective map.
Proof (i) We prove this claim in full detail to provide a model for
arguments of this type. We have to check the axioms listed in (3.5). First,
we need 0 G Ker(0). In M, 0 + 0 = 0. Thus, in N, 6(0) = 6(0) + 6(0). But
6(0) has a negative in TV, so we get the equations
0 = 6>(0) - 6(0)
= (0(0)+ 0(0)) -0(0)
= 6(0)+ (6(0) -6(0))
= 6(0) +0
= 0(0).
6(rm) = r ■ 9(m)
= r-0
= 0,
so that rm G Kei(6).
(ii) ■$=: Suppose that 6(m) = 6(n) for elements m, n of M. Then 6(m - n) — 0,
so m - n G Kev(6), giving m - n — 0 and m = n a s required.
60 Chapter 4. Homomorphisms
\° / \° / \l )
is the standard basis of Fp. A direct calculation, to be spelt out in a
wider context in the next chapter, shows that the vectors { T e i , . . . ,Te p }
are simply the columns of T, so that lm(0) is the subspace of Fn spanned
by the columns of T. This space is familiarly known as the column space
of T.
The dimension of lm(0) is called the rank of 0 or of T, and written
rank0 or rank(T).
The rank and nullity are connected by the following well known result,
which is called variously the Rank and Nullity Theorem or Kernel and Image
Theorem:
rank(0) + null(0) = p.
4.11. Some calculations 61
Ker(<7(3))= {0,2,4}
and
Im(<7(3)) = {0,3}.
(ii) Let R be arbitrary. For a fixed a; in M, r ( i ) : fi —\ M is given by
T(X)T = rx. The image Im(r(a;)) is simply the cyclic submodule Rx of M
generated by x (as in section 3.8), so that T(X) is surjective precisely when
M is cyclic with x as a generator.
The kernel of T(X) is called the annihilator Ann(:r) of x:
Ann(x) = {r G R \ rx = 0}.
Ann (a;) is a left ideal of R that will play an important part in our discussions
in later chapters.
(iii) Finally, we take M to be the module of section 4.8, which is given
by the matrix A = I n „ I over an arbitrary field F, and we compute
the kernel and image of each F[X]-homomorphism from M to itself. Recall
' t t
that such a homomorphism is given by a 2 x 2 matrix T - ' n
0 in
x
\ - "2
with tn and £12 arbitrary, so that, for v = f ) G F , we have Tv =
tnx + ti2y
tny
There are two approaches, one by direct assault and one that uses a
little subtlety. We give both for illustration. The direct calculation of the
' 0
kernel, those v with Tv = I 1, falls into three cases.
If i n / 0, then tnV = 0 gives y = 0 and then x = 0, so that Ker(T) = 0.
62 Chapter 4. Homomorphisms
tnx + t12y \ _ / a
tuy ) \ b
If t n 7^ 0, we can solve for y and x in succession, so T is then surjective.
If tn = 0, we can only obtain those w with 6 = 0. If t 12 ^ 0, we can always
solve the equation ti2y = a, which shows that Im(T) = I _ I . Finally,
T = 0 implies Im(T) = 0.
or
Ker(T) = L and Im(T) = L,
or
Ker(T) = 0 and Im(T) = M.
The first combination corresponds only to T = 0. For the second, we must
have t n = 0, by direct observation, and ti 2 ^ 0. Thus if both tu and t i 2
are nonzero, we must be in the third case.
4.12 Isomorphisms
In our study of the structure of modules, it will be important to know when
two superficially different modules are in essence the same. This is the case
when there is an isomorphism between them.
4.12. Isomorphisms 63
4.12.1 Proposition
Let R be a ring, let M and TV be R-modules, and let 9 : M —> TV be an
R-module homomorphism. Then the following statements are equivalent,
(i) 9 is an isomorphism,
(ii) 9 is invertible.
(Hi) Ker(0) = 0 and lm(0) = TV.
4.13.1 P r o p o s i t i o n
Let R be a ring, let M and N be left R-modules, and let 9 : M —>■ N be
a surjective R-module homomorphism. Then the following assertions hold.
(i) Let L be a submodule of M with Ker(0) C L. Then
9*9*{L) = L.
(ii) Let P be a submodule of N. Then
9,9*(P) = P.
(Hi) The maps 9, and 6* are mutually inverse bisections between
the set of submodules L of M that contain Ker(0)
and
the set of submodules P of N.
Explicitly,
L^6,(L)
and
P^9*(P).
Proof
(i) It is obvious that L C 9*9*(L). To prove equality, take any m €
6*9,(L). Then 9(m) = 9(1) for some I € L, so that
m - I = k G Ker(0) C L.
Thus m € L.
Exercises 65
(ii) It is clear from the definition that 0*6* (P) C P. But if p £ P, then
p = 9{m) for some m £ M, as 0 is surjective. Thus m € 0*{P),
again by the definition, so we have equality.
The final assertion is now obvious. D
Exercises
4.1 Let R be a commutative ring and let M and N be R-modules. Let
Hom(M, N) be the set of all .R-module homomorphisms 9 : M -> N'.
Show that the addition defined in section 4.2 makes Hom(M, N) into
an additive group, with zero element the zero map, and with — 9
defined by -9(m) = -(0(m)) for all m £ M.
For any r £ R, define r9 by
(rO){m) =r(0{m))
for all m. Verify that r0 £ Hom(M, TV) and hence that Hom(M, N)
is also an .R-module.
4.2 Let R be a ring, let M, N and P be .R-modules and suppose that
0,i> e Hom(M, N) and <j>,p £ Hom(iV, P).
Verify that
(f,(9 + i>) = 00 + 4>ip
and that
(cf) + p)9 = 09 + P9.
u{<t>9) = (w</>)9.
4.3 Combine the results from the above exercises, show that Hom(M, M)
is a ring.
Aside: this ring is called the endomorphism ring of M. Usually, the
endomorphism ring of a module is noncommutative (see Exercise 5.9).
Thus endomorphism rings do not play an explicit role in this text, in
marked contrast to their fundamental importance in ring theory in
general.
4.4 Let R be a commutative ring, let M be an R-module and write S =
Hom(M,M). For a in R, let
be the multiplication map defined in section 4.3. Verify that for any
elements a,b 6 R,
a(a + b) = a (a) + a(b)
and
a(ab) = cr(a)a(b)
and that
o-(lfi) = I s = idM-
Aside: in general, a map a from a ring fitoa ring S that satisfies the
above equalities is, by definition, a ring homomorphism. A bijective
ring homomorphism is a called a ring isomorphism. Ring homomor
phisms play only a minor role in these notes, although we have used
them implicitly in our discussion of residue rings of polynomials in
section 2.12 - when we regard a scalar k G F as the same as its image
k G F[X]/F[X]f, we are using the fact that the map k —> k is an
injective ring homomorphism.
4.5 Take M = R in the preceding exercise. Show that a(a) is an injective
homomorphism on R for all nonzero a € R «=^> R is a domain.
Show also that <r(a) is surjective for all nonzero a in R if and only
if R is a field.
4.6 Let R be a commutative ring and let M be an .R-module. For each x
in M, let
T(X) : M -> Hom(i?, M), r(a;)r = rx,
be as in section 4.3. Show that r is an isomorphism of .R-modules.
4.7 Let F be a field, let M be F p made into an F[X]-module through a
p x p matrix A, and let N be F made into an F[X]-module through
a constant A (so we are in the reverse situation to the example in
section 4.7).
Show that the F[X]-module homomorphisms from M to TV are
given by the row vectors w of length p with wA = Aw.
Note: a square matrix A has two sets of eigenvectors, those satisfying
Aw = Xw and those satisfying vA = Av. These are the right and left
eigenvectors of A, respectively. As we nearly always regard a matrix
as a left operator, we use the term "eigenvector" to mean a right
eigenvector.
4.8 Let F be a field and let M and N be F[X]-modules given by the
F-linear transformations a and /? respectively. Let M(h) denote the
F[X]-module given by ah for h > 1 and define N(h) similarly.
Show that if 6 : M —> N is an F[X]-module homomorphism, then
9 : M{h) —► N(h) is also a F[X]-module homomorphism for all h > 1.
Exercises 67
4.9 Here are 5 modules over R[X], listed with the 2 x 2 matrices which
define them.
0 1
L :A =
1 0
0 -1
M :B = I " ■■
" 1 (see the example in 3.12)
1 0
1 1
N :C = I ) (see the example in 3.11)
0 1
0 1
P :D = ( n j (see the example in 4.8)
0 0
0 1
Q :E =
-1 0
Free Modules
69
70 Chapter 5. Free Modules
\n )
the ring R. The set .R-module structure on Rk is given by the expected
rules for addition and scalar multiplication: if
/ si \
1"2 Si
m: and n are in R ,
\rk J \sk )
then
( ri + si \
r 2 + s2
m +n
and for r m. R,
( rri \
rr2
rm =
V rrk )
A routine verification confirms that Rk is a left R-module. When the ring
of coefficients is a field F, Fh is familiar to us as the standard column space
of dimension k.
The standard basis of Rk is the set
/ 1 \ /o\ ( 0 \
ei = , ej — ■i efc =
\0J \0 J \ i /
where, for j = 1 , . . . , k, ej has the entry 1 in the j-th place and zeroes
elsewhere.
5.2. Free modules in general 71
m = 7-i&! + . . . + rkbk-
r-ybi H \-rkbk =0
holds only if
n = • • ■ = r;. = o.
Then B is a basis of M if it is linearly independent over R and it
generates M as an .R-module.
A free R-module is denned to be an R-module M that has a basis.
The number of elements in the basis is called the rank of M, and written
rank(M).
It is clear that the standard basis of Rk, as defined in the preceding
section, is actually a basis of the standard free module Rk, and consequently
Rfc is indeed free, of rank k. By our conventions, the zero module 0 = R°
is free of rank 0 since its basis is the empty set.
Before we commence a detailed discussion of bases, here are some points
to bear in mind.
• When the ring of scalars is a field F, a finitely generated F-module is the
same thing as a finite dimensional vector space over F. By elementary
linear algebra, such a vector space V always has a basis, and the rank of
V is usually referred to as the dimension of V.
72 Chapter 5. Free Modules
• Two basic results in elementary linear algebra are that any linearly inde
pendent subset of a finite dimensional vector space V over a field F can
be extended to a basis of V, and that any generating set of V contains a
basis of V.
These results do not hold for modules over more general rings. For
example, the subset {2} of Z is linearly independent but Z has no basis
of the form {2, o , . . . } , whatever the choice of a , . . . in Z. The set {2,3}
generates Z as a Z-module, since 1 = 2 - 2 - 3 , but no subset of {2,3} is
a basis.
• A consequence of the results quoted above is that every finite dimensional
vector space over a field F is a free F-module. When the coefficient ring
R is not a field, we expect to find .R-modules that are not free (Exercise
5.1). For example, the Z-modules Z m , m > 0 (1.11) contain no linearly
independent subsets, since mx = 0 for any x e Z m .
• Warning! Some authors define bases in a different way which allows the
possibility that Z m has a basis as a Z-module; however, the definition of
a free module must then be altered.
• Whether or not a module is free depends on the ring of scalars, as do
the concepts of linear independence and generation. Consider the residue
ring Z p where p is a prime. This is a field (1.11.2) and so free of rank 1
as a Zp-module, but it is not free as a Z-module.
• Another illustration of the same type is provided by the standard vector
space Fk over a field F, made into an F[X]-module M with X acting as 0.
Then M is not free as an F[X]-module as it has no linearly independent
subsets over F[X].
• According to our definition, the rank of a free module depends on the
basis B. We shall see soon (Theorem 5.5.2) that the rank is in fact
independent of the choice of basis (at least, for the rings of most inter-
set to us in these notes). However, there are rings for which the rank
of a free module can vary with the choice of basis - see section 2.3 of
[B & K: IRM].
• Our definition of a free module requires that the rank is finite. Since
free modules of infinite rank play only a minor role in this text, we
have relegated their definition to Exercise 7.10. The extension of the
definitions and results of this chapter to modules with infinite ranks is
discussed in [B & K: IRM], (2.2.13).
5.2.1 Lemma
Let R be a ring, let M be a left R-module, and let B = {&i,..., 6fe} be
a subset of M.
5.3. A running example 73
0 = nbi H \-rkbk
0 = Oh + 1- 0bk
n + 2r 2 = 0
- r i + ar2 = 0
n + 2r 2 = 1
—r\ + ar 2 = 0
e\ = —b\ + 62
e2 = — 2£>i + 62
*=(sH
we have
x = ( - i i - 2x2)h + (xi + x2)b2.
This confirms that, for a = —1, B is a basis of Z 2 ; the reader is recom
mended to make the corresponding calculation for a = —3.
5.4.1 T h e o r e m
Let R be a ring and let M be a left R-module. Then there is a bijective
correspondence between
(i) the set of all R-module isomorphisms 6 : Rk —> M
and
(ii) the set of all bases B = {b\ ..., bk} of M.
5.4. Bases and isomorphisms 75
m = ^(x)
= 6>(a;iei) + ••■ +6{xkek)
= xi9(ei)-\ \-xk9(ek)
= xih-\ \-xkbk,
*-(_! u= .;
of Z 2 which we constructed in section 5.3. The theorem predicts that there
is a corresponding isomorphism 9 : Z 2 -> Z 2 with 9{e\) = b\ and #(e 2 ) = b2-
76 Chapter 5. Free Modules
- x i - 2x2
tp(x)
X
Xi+ X2
5.5.1 Lemma
Let R be a commutative domain with field of fractions Q. Then the
following statements hold.
(i) If qi = Ti/Si, i = 1 , . . . ,k is any finite set of elements in Q, there is
an element s € R and elements ai € R with qi = Oj/s for all i.
Note: s is called a common denominator of q\,... ,qk, and this
rewriting process is known as "placing over a common denomina
tor".
(ii) Rk C Qk for anyk>\.
(Hi) If v £ Qk, then v = (l/s)m for some s 6 R and m € Rk.
(iv) If B is an R-basis of Rk, then B is also a Q-basis of Qk.
(91 \
n\
V = = ((V») = (l/s)m
i
\9k ) V ak )
5.6. Change of basis 77
5.5.2 T h e o r e m
Let R be a domain and let M be a free R-module. Then any two bases
of M have the same number of elements.
Proof Suppose that M has two bases, one with h elements and one with
k elements. By the previous result, there are isomorphisms 6 : Rh —5- M
and <)> : Rk —» M, and hence an isomorphism 6~14> : Rk —> Rh. Thus the
standard free module Rh itself has a basis B with k elements, by Theorem
5.4.1.
But the dimension of a vector space over the field of fractions Q is
unique, and, as B is also a basis of Q , we have h = k. □
Remark: the rank of a free .R-module is unique for any commutative ring
R (Exercise 5.2).
bk = O61 + 06 2 + • • ■ + l i t
which gives t h e identities
k f 0 if h ^
(5.5)
\ lifh =
5.7. Coordinates 79
PB,CPC,B = I, (5.6)
PC,BPB,C = I (5.7)
also, so that PC,B and PB,C are mutually inverse matrices over R.
We illustrate these calculations with the basis B = {61,62} of Z 2 ,
♦.-(-DM.;)-
which we considered in sections 5.3 and 5.4, taking C = E, the standard
basis.
It is clear that
M-l -0-
that is, the columns of Pc,s are simply the vectors 61,62 themselves.
To compute PB,C, recall from section 5.3 that
ei = -61 + 62
e2 = -26j + 62;
thus
Ml "0-
If we are given a basis B of a free module and an invertible matrix Q
of the correct size, we can construct a new basis C so that Q = PB,C but,
before we do so, it is convenient to discuss coordinates.
5.7 Coordinates
Given a free .R-module M with basis B = {&i,...,6fc}, we know from
Lemma 5.2.1 that for m e M, there is a unique set of coefficients { r i , . . . , r^}
in i? with
m = ri&iH +r fc 6 fc .
80 Chapter 5. Free Modules
The element
MB = I i I e Bk
m = ri(pnci H hPfciCfc)
+ V rk(j>ikci H 1- PfcfcCfc)
= {pun H hpifcrfc)ci
+ r {pkin + r PkkTk)ck
which can be summarized as
Notice that when E is the standard basis of the standard free module
Rk, we have
k
(X)E = x for all x € R .
\ xi + X
Xi x22 J
for our illustrative basis B, while (with C = E),
«■* - - 1 1 ))
by the computation in section 5.6. An easy verification confirms Formula
5.8.
First, notice that the identities in Eq. (5.1) still hold - this is confirmed
by substituting for Ci,...,Cfc in the right-hand side of the equation and
using the fact that P is given to be the inverse of Q. It follows that C
generates M, for any m e M has the form
m T-I&H \-rkbk,
and so can be expressed in terms of c\,..., ck by substitution.
Suppose next that 0 = siCi + ■ ■ ■ + SkCk for some scalars 8\,..., s& in
R. Substituting for each c, in terms of the bi's and calculating coefficients
as in section 5.7 above, we find that
0 ='- (0)B = Qs
where
s-
\ sk
Then s = PQs = 0, so that C is linearly independent and hence a basis for
M.
Notice that Q = PB,c and P = PC,B-
V tn ti2 Uk )
82 Chapter 5. Free Modules
5.9.1 Proposition
Let R be a commutative domain and let M, N and P be free R-modules,
with bases B, C and D respectively. Then the following hold.
(i) {idM)s,B = I, where id^ is the identity map on M and I is a kx k
identity matrix, k = rank(M).
(ii) If 0 : M —¥ N and 4>: N —»■ P are R-module homomorphisms, then
{<fi)D,c(e)c,B = (4>0)D,B-
4>6(bj) = 4>(J2UjCi\
i
= y2uj4>{ci)
i=\
5.10. Illustration: the standard case 83
fc 1 IIL
i=\
'\J2Sh idh I
\h=l
m / I \
1 dh,
h=\ \i=\ J
5.9.2 Corollary
Let M and N be free R-modules, with bases B and C respectively, and
let 9 : M —> N be an R-module homomorphism.
Then 6 is an isomorphism if and only if (0)c,B is an invertible matrix,
in which case {{0)C,B)~ = ( # - 1 ) B C-
(6)c,B(0-1)B,c = (idN)c,c=I
and
{6-1)B,G{6)C,B = (idM)B,B = I.
4=: If (9)c,B has an inverse S = (shi), take <j>: N —> M to be the homomor
phism with matrix S. Then both products 9(/> and <j>6 have as matrix the
identity matrix, so they are both identity maps, on N and M respectively.
Thus (j> = 9~1. U
T
= (9)E(i),E(k)
84 Chapter 5. Free Modules
which is an / x k matrix whose i-th column is simply the image 0(e;) of the
i-th standard basis vector e^ in E(k).
Thus for many purposes the homomorphism 9 can be viewed as being
effectively the same as the matrix T it defines. However, this identification
of a homomorphism and a matrix is dependent on the fact that we use the
standard bases, and so it can be misleading when we use nonstandard bases
of standard free modules.
When the ring of scalars is a field F and the free modules are the
standard vector spaces Fk and Fl with their standard bases, we recover
the correspondence between F-linear transformations and matrices that is
familiar from elementary linear algebra, as promised in sections 4.6 and
4.10.
TC,B = (thi)
and
TC,B> = (t'gj)-
We claim that these matrices are related by the formula
where Pc,c aud PB,B> a r e change of basis matrices. There are two ap
proaches to the verification of this formula. One is simply to expand the
matrix on the right and check that we have the desired equality. Although
this method is elementary, it does require a lot of detailed calculation. A
more sophisticated approach is to use the calculations that we have already
performed, together with a nice observation.
If we compare the formulas given in Eqs. (5.1) and (5.9), we see that the
change of basis matrix PB,B' is just the basis of the identity transformation
on M with respect to the pair of bases B',B, so that
Likewise
Pc,c = c(idN)c-
Since
6 = idjv • 0 ■ id,M,
we obtain the relation 5.13 immediately from the product formula for ma
trices of homomorphisms that we obtained in Proposition 5.13.
>4ii = ( - l ) < + i d e t ( m y ) .
We can then take as our working definition the formula
that is, expansion from the first row. For example, when k = 2, An = a22
and A\i = —021, so we recover the familiar formula.
We assume the basic properties of the determinant, which we list below
for future reference. These results hold for matrices with entries in any
commutative ring. Proofs can be found in Chapter 7 of [Cohn 1].
Det 1: Let A be a k x k matrix over a commutative ring R. Then det (A) E R.
Det 2: If B is also k x k matrix over R, then
for row vectors T>i and Cj. Write B for the matrix which is formed from
A by replacing row a^ by b*, and define C similarly. Then
det(A) = det(B) + det(C).
There is a corresp onding result for columns.
Det 7: If
/on 0 ... 0 0 \
021 ^22 • •• 0 0
A=
flfc-1,1 a/c-i,2 • ■• a-k-i,k-i 0
\ Ofel afc2 ■ ■ afc,fc-i afcfc /
is a lower triangular matrix, then
det(yl) = ana22 ■ ■a-kk-
In particular, det(7) = 1.
Det 8: The following relations hold:
det(A) for h=i
ahiAn + • ■ • +dhkAik
-{ 0 for
and
det (A) for i = j
Auaij + ■ ■ ■ + Akidkj
-{ 0 for
In the case h = i, the first relation tells us that the determinant can
be expanded from row h for any h = 1 , . . . , k. Similarly, for i = j , the
second tells us that we can expand from the j - t h column.
Det 9: The adjoint of A is adj(A) = (Aij)T, the transpose of the matrix of
cofactors of A. Then the formulas of Det 8 can be re-interpreted as the
matrix product formulas
A ■ adj(A) = det(A)I = adj(A) • A
where the middle term is simply the diagonal matrix with all diagonal
terms equal to det(A).
5.12. Determinants and invertible matrices 87
This last property leads to the invertibility criterion that we have been
seeking.
5.12.1 Theorem
Let A be a k x k matrix over a commutative ring R. Then A has an
inverse with entries in R if and only i/det(,4) is a unit in R.
Furthermore, if A is invertible, then
'-*.-(-1.1)
must be invertible, by the results of section 5.6. On the other hand, if P is
invertible, then B is a basis by section 5.8.
Now det(P) = a + 2, which is a unit in Z when a = - 1 or a = - 3 , thus
confirming the calculation made in section 5.3.
88 Chapter 5. Free Modules
Exercises
5.1 Let R be a commutative ring and suppose that the nonzero element
r € R is not a unit of R. Show that the i?-module R/Rr is not a free
i?-module. (Hence a ring which is not a field has non-free modules.)
5.2 (a) Let P be an invertible matrix over a commutative domain R. Using
Theorem 5.5.2, show that P must be a square matrix.
(b) Let R now be any commutative ring. Show that if & : R —> R
is an isomorphism, then there is an invertible I x k matrix T with
Tv = 9(v) for all v e Rk.
(c) If T is not square, add zero rows or columns to obtain a square
matrix, and so obtain a contradiction, using Det 8. Deduce that
Theorem 5.5.2 holds for any commutative ring.
5.3 Let P be k x k matrix over a commutative domain R. Show that P
is invertible if and only if the columns of P form a basis B of Rk, in
which case P = PE,B-
5.4 Let B and C be bases of a free i?-module M. Use the fact that
{idM)c,B = Pc,B
(m)c = Pc,B{m)B
(9(m))c = (6)c,B(m)B.
5.6 Let
and
- - U h - -0)}
D
2
- * - -i -*=U))}
be bases of Z (see section 5.3).
Compute PD,B-
5.7 Find all values of a in the Gaussian integers Z[i] for which the elements
b\ = [ I and 62 = I j form a basis of Z[i] 2 .
Exercises 89
5.8 Let g(X) be a polynomial over 11 Find all f(X) in R[X] for which
the elements b\ = ( I and hi " ( fun i formabasisofR x 2
> /""""- [ ] -
5.9 Let R be a commutative ring and let E be the standard basis of Rn.
For each 0 in Eom(Rn,Rn), put a(0) = E{0)E- Show that a is a ring
s
isomorphism from Hom(/? n , Rn) to the ring Mn(R) of n x n matrices
over .R.
Let M be any free R-module with rank(M) > 1. Deduce that the
ring Hom(M, M) is not commutative.
5.10 Let
0 . 0 0 \
021 a22 . 0 0
A =
fl
fc-i,i afc- 1,2 ■ • ajt-i,fe-i 0
\ ajti afc2 • afc,fc-i «fcfc /
be a lower triangular matrix over a commutative ring R.
Show that A is invertible if and only if all the diagonal terms
011,022, • • ■ ,afcfc a r e units in R. Hint: Det 7.
Show further that if A is invertible, the inverse of A is also lower
triangular.
Chapter 6
91
92 Chapter 6. Quotient Modules and Cyclic Modules
m = n <=$• m — n € L.
m = m + L = {m + I | I G L};
m + n = m + n for m, n G M/L,
More routine verifications, very similar to those made in the proof of Propo
sition 1.10.2, show that these operations are well-defined and make M/L
into a left .R-module, with zero element 0.
6.2.1 L e m m a
Let L be a submodule of M and let w : M —> M/L be the canonical
homomorphism. Then
Ker(7r) = L
6.3. Induced hoznomorphisms 93
Proof
m e Ker(7r) <=> m = 0
«=> m - 0€ L
<=> m e L.
m e Ker(0) «=> 9m = 0
<=> m£ Ker(0).
6.4 Cyclic m o d u l e s
As an application of the First Isomorphism Theorem, we show how to
describe cyclic modules and their submodules.
Recall from section 3.8 that an .R-module M is cyclic if M = Rx for
some element x in M. This is equivalent to saying that the multiplication
homomorphism of section 4.3,
Ker(r) = {r £ R | rx = 0},
which is the annihilator ideal Ann(x) of x (sections 3.5 and 4.11). Write
I = Ker(r). Then, by the First Isomorphism Theorem, f is an isomorphism
from R/I to M.
In the other direction, suppose we are given a left ideal / of R. Then
the quotient module R/I is cyclic with generator 1, since r = r ■ 1 for any
reR.
The left ideal I that we have associated with a cyclic module M depends
on the generator x of M that we have chosen. To complete the classification
of cyclic modules, we need to show that I depends only on M. In fact, we
prove an apparently stronger result, that the ideal I is uniquely determined
by the isomorphism class of the cyclic module. Suppose that M and N are
cyclic left .R-modules and that there is an .R-module isomorphism
a : M -> N.
0 = x-j(T) = j(x),
6.4.1 Theorem
Let R be a ring and let M be a cyclic left R-module. Then there is a
left ideal I of R such that M = R/I as an R-module.
If N is also a cyclic left R-module with N = R/J for some left ideal J,
then M = N as an R-module if and only if I = J.
In particular, the left ideal I is uniquely determined by the cyclic module
M. □
96 Chapter 6. Quotient Modules and Cyclic Modules
Remarks
(i) These results include the two extreme cases of cyclic modules, which we
have not mentioned explicitly until now. If / = 0, the zero ideal, then
R/I = R. U I = R, then R/I = 0, the zero module.
(ii) Suppose that ideal / is two-sided (as is always the case when R is com
mutative). Then the quotient module R/I is, in essence, the same as
the residue ring R/I of section 1.10 (which is the reason that we can use
the same notation for these two constructions). Both have the same rule
of addition, and the scalar multiplication is related to the residue ring
multiplication by the formula r -x = Tx.
(iii) As we have seen in our discussion of matrix actions on vector spaces, a
single additive group can usually be viewed as a module in many different
ways. Any module whose underlying group A is cyclic is itself necessarily
cyclic. Then A must be isomorphic to Z or to Z p for some prime p.
The converse is far from true. For example, the ring of Gaussian
integers Z[t] is cyclic as a module over itself, but not as a module over
Z since Z[i] = Z 2 . Other examples are provided by the C[X]-modules of
Exercises 3.6 and 3.7; both these modules are cyclic, but the underlying
vector space C 3 is not cyclic over C, nor does it give a cyclic C[X]-module
when X acts as 0.
6.5 S u b m o d u l e s of cyclic m o d u l e s
We now combine the description of cyclic modules given above with the
submodule correspondence that we obtained in section 4.13 to find the
submodules of a cyclic module.
Let R be an arbitrary ring. We make repeated use of the observation
that the left ideals of R are precisely the i?-submodules of the left regular
i?-module R. Choose a left ideal I of R and write n : R —» R/I for the
canonical homomorphism of left .R-modules. If P C R/I is an i?-submodule
of R/I, then the inverse image of P is
6.5.1 P r o p o s i t i o n
Let I be a left ideal I of R and let M = R/I be the cyclic left R-module
defined by I. Then there is a bijective correspondence between
6.5. Submodules of cyclic modules 97
6.5.2 Theorem
Let R be a Euclidean domain and let R/I be a cyclic R-module, where
the ideal I is neither 0 nor R. Then the following statements are true.
(i) I = Ra, where a has a standard factorization of the form
n(l) n(Jfe)
Proof (i) By Theorem 2.5.3, the ideal I is principal, say / = Ra' where
l nil) n(k) .,
a = upj v ' .. .pk , u a unit,
is a standard factorization of a' (section 2.9), and a = u _ 1 a ' is also a
generator of / (Lemma 2.5.3), with the desired standard form,
(ii) By the preceding result, the submodules of R/Ra are given by the
ideals H of R with
Ra C H C R.
98 Chapter 6. Quotient Modules and Cyclic Modules
r £ Ker(0) rd = 0
rd € -Ra
r € fid'
so that Ker(0) = Rd'. The First Isomorphism Theorem (6.3.2) now shows
that R/Rd' 3! fid/ito. □
pM qM
1 e ei en~l
en = -fo-he fn-ie71'1.
X 1 =e
X e = e2
(6.1)
£ n --2 -n —1
X
X e™"-1 = - / o - 1 --he- _ / 2 £ 2 /n-ie n - 1
100 Chapter 6. Quotient Modules and Cyclic Modules
Thus, by the definition in section 5.9, the matrix of the linear transfor
mation corresponding to X is
(o 0 • • 0 0 -/o \
1 0 ■ • 0 0 -h
0 1 ■ • 0 0 -h (6.2)
C(f) =
0 0 ■ ■ 1 0 — fn-2
0 • • 0 1 -/n-1 j
The matrix C(f) is called the companion matrix of the polynomial / or the
rational canonical block matrix associated to / , since it is a typical building
block for the rational canonical matrices that we encounter later.
Here are some special cases. A linear polynomial / = X — a, has
C ( / ) = ( a )i a 1 x 1 matrix. Since a e F is arbitrary, we see that any
method of turning F into an F[X]-module must result in a cyclic module,
which fact is obvious anyway.
For a quadratic polynomial / = X2 + aX + 6, we have
C(f)
6.7.1 T h e o r e m
Let B be an n x n matrix over a field F, and let M be Fn made into
an F[X]-module with X acting as B. Then the following assertions are
equivalent.
(i) M is isomorphic to a cyclic F[X}-module F[X}/F\X]f for some
monic polynomial f € F[X].
(ii) There is an invertible n x n matrix T so that TBT~l is a rational
canonical block matrix C.
When these assertions hold, C = C(f) for a unique monic polynomial f in
F[X].
Proof
Before we can get started, we need to set up some notation. Given B,
the action of X as B on Fn defines an F-linear transformation
(ii) => (i): Given T and C, define 6 by the relation (6)Z,E = T, and let /
be the monic polynomial with C = C(f). Reversing the above argument,
we see that 6 is an F[X]-module homomorphism from M to F[X]/F[X]f.
Finally, we show that the monic polynomial / is uniquely determined by
M. If M £ F[X]/F[X]f and also M S F[X]/F[X]fc, then F [ X ] / F [ X ] / a
F[X]/F[X]h, so that F [ X ] / = F[X]h by Theorem 6.4.1 and therefore
f = h. □
Exercises
6.1 Let R be a commutative ring and let M be an i?-module. Recall
from Exercise 4.6 that every i?-module homomorphism p : R —> M is
determined uniquely by x = p(l) 6 M; then p(r) = rx for all r G R.
Let a be a fixed element of R. Show that
M(a) = {x £ M | ax = 0}
is a submodule of M.
Show that if £ : R/Ra -> M is an .R-module homomorphism, then
C, = p where the element x corresponding to p belongs to M(a). Show
conversely that if x 6 M(a), then x gives rise to a homomorphism
from R/Ra to M.
Exercises 103
K + L = {k + l | ke K, leL}.
K/(Kr\L) S {K + L)/L.
(M/K)/(L/K) £ M/L.
0 0 1 — fn-2
V0 0 0
be a rational canonical block matrix, and put / /o + / i X + ■ • • +
fn-iXn~l + Xn, so that C = C(f).
Let I be an n x n identity matrix. Show that
det{XI -A)=f.
Remark. Thus / is the characteristic polynomial of C, which we meet
again in section 9.7.
Hint: use row operations - see section 5.12.
6.6 Let R be a Euclidean domain and let l,p, q be distinct irreducible
elements of R. Draw up diagrams that illustrate the submodules of
R/Ra (as in section 6.5) when
(a) a = p;
(b) a = p2;
(c) a = Ipq;
(d) a = p2q;
(e) a = p2q2.
s 6.7 Let R be any ring. A composition series for a left .R-module M is a
finite ascending chain
0 = M 0 C Mi c • • • C Mfc_x c Mk = M
in which each quotient Mi/Mi_i, i = l,...,k, is a simple left R-
module.
Suppose that R is a Euclidean domain and that M = R/Ra is
cyclic, a / 0. Find a composition series for M , and show that the
set {Mi/Mi-i | i = l,...,k} corresponds bijectively to the set of
irreducible factors of a (counting multiplicities).
Does R itself have a composition series?
Remark. If a module has a composition series, then the set of simple
quotient modules associated to the series is essentially unique. This
classical result is the Jordan-Holder Theorem, a proof of which can
be found in [B & K: IRM] (4.1.10).
Exercises 105
0 C J i C • - • C Ife C 4 + 1 C--- C R
i2 = j 2
= k2 = — 1 and ij = k,
X2 - T{v)X + N{v) = 0.
107
108 Chapter 7. Direct Sums of Modules
IDSM 1: L + N = M;
IDSM 2: L n TV = 0.
The notation M = L © TV indicates that M is the internal direct sum of its
submodules L, TV, which are then called the components or summands of
M. We also say that TV is the complement of L in M, and vice versa.
When we have expressed a module M as a direct sum of two (or more)
components, we sometimes say that we have decomposed M or that we have
found a decomposition of M. Before giving any examples, we reformulate
the definition in a useful way.
7.1.1 Proposition
Let L and TV be submodules of a left R-module M. Then the following
assertions are equivalent.
(i) M = L®N.
(ii) Let m g M. Then there are unique elements I € L and n € TV with
m = I + n.
Proof
(i) =>• (ii): Suppose that m € M is given. Since M = L + N, we have
m = I + n for some I G L, n € N. If also m = V + n! with /' € L, n' 6 N,
then l — l' = n' — n belongs to LC\N, which is 0. Thus I and n are uniquely
determined by m.
(ii) => (i): The fact that there are elements / and n with m = I + n for
each m in M shows that M = L + N. Suppose that x 6 LflJV. Then
£ = x + 0 with i e i and 0 s TV, and also a: = 0 + x with 0 € L and x € TV.
By uniqueness, we must have x = 0. □
The above comments hold when the field F is replaced by any ring
R.
(iv) The previous example illustrates that a module can be expressed as a
direct sum in many ways, since a vector space has many bases.
It also shows that the choice of one component of a direct sum need
not determine the other. To see this, consider the bases of the form
{ e i , / ( a ) } with / ( a ) = I °. ), where a <E F is arbitrary. Then Ff(a) ^
Ff(b) if a ^ b, and V = Fex © Ff(a) for every choice of a.
(v) Let_M = Z 6 , considered as a Z-module. Then 2M = {0,2,4} and 3M =
{0, 3} are submodules of M with 2M n 3M = 0. The fact that M =
2M + 3M follows either by direct calculation or, more intellectually, from
the identity 1 = 2 - 2 - 3 .
In contrast to the previous example, 2M and 3M are the only non-
trivial summands of M.
This example foreshadows a general technique for constructing de
compositions of modules over a Euclidean domain.
(vi) Let F be a field, let A — f I be a 2 x 2 diagonal matrix over F,
and let M be the F[X]-module defined by X acting as A on F2. Then the
subspaces L = Fei and N = Fe^ are F[X]-submodules of M, the action
of X on L being given by multiplication by b and on N by multiplication
by c. We have M = L © N.
Notice that we already know that M is the direct sum of L and N as
a vector space over the field F; the real meat of this example lies in the
fact that L and N are invariant under the action of X.
(vii) More generally, suppose that ^ 4 = 1 n ^ l i s a f c x f c block matrix
over F, with diagonal blocks of sizes r x r and s x s respectively. (Hence
r + s = k.)
Let M be the .F[X]-module defined by X acting as A on Fk, and let
L = Fe\ + - ■ -+Fer and TV = Fe r + iH hFek, where, as usual, e\,..., efc
is the standard basis of Fk. Both L and N are .F[X]-submodules of M,
the action of X on L being given by the matrix B and on N by C. Then
M = L © N. (A more general version of this construction is given in
section 7.5 below.)
L+N
LDN
On the other hand, when M is the direct sum of L and N, the equations
Lf) N = 0 and L + N = M lead to the simpler diagram
M •
L • N
0 •
7.3. Indecomposable modules 111
(If one of the submodules L, N is zero, the diagrams become even simpler.)
7.3.1 Theorem
Let p be an irreducible element in a Euclidean domain R. Then the
cyclic R-module R/Rpn is indecomposable for any integer n > 1.
IDSMkl: Li + --- + Lh = M;
IDSMk 2: Li n (Li + ■ • • + U-i + Li+1 + h Lk) = 0 for i = 1 , . . . , k.
The notation for such an internal direct sum is M = L\ © • • • © Lfc. The
submodules Li, i = 1,... ,k are the components or summands of M, and
the complement of L* is the submodule
The order of the terms is unimportant, and we allow the possibility that
some components are the zero module.
It is convenient to allow the trivial cases k = 0, where M = 0, and
k = I, where M = L\. For k = 2 we regain the definition of the direct sum
of two submodules.
Notice also that if M = L\ © • • • © Lfc, then M = Li® Li for each i.
Internal direct sums with many components will occur frequently later
in these notes. An immediate example is provided by the standard left free
module Rk. For any basis {bi,..., bt} of Rk,
Rk = Rh © ■ • • © Rbk
by Lemma 5.2.1.
The proof of the following useful but straightforward extension of Propo
sition 7.1.1 is left to the reader.
7.4.1 Proposition
Let L i , . . . , Lfc be submodules of a left R-module M. Then the following
assertions are equivalent.
(i) M = Li © • • • © Lfc;
(ii) Let m € M. Then there are unique elements
with
m = li + ■ ■■ +/fc.
□
7.5. Block diagonal actions 113
\ 0 0 Dk J
be a block diagonal matrix over F, with k blocks on the diagonal, and
suppose that D is an s x s matrix. The action of D on F" defines an F[X\-
module M which can be expressed very naturally as an internal direct sum.
Since this type of decomposition is very important in future applications,
particularly in Chapter 13, we give the full details, although they appear
rather gruesome at first sight.
As a first illustration, we consider the case that
/ di 0 \
0 d2
D =
\ 0 0 ds J
is an s x s diagonal matrix over F (so that k = s).
Put Li = Fei for i = 1 , . . . , s. As mentioned in the previous section, we
already know that M = L\ © • • • © Ls as a vector space over F. However,
each Li is also an F[JV]-submodule of M, with X acting as di, and so we
have a decomposition of M as an internal direct sum of F[X]-submodules.
For a general block diagonal matrix D as above, we have to overcome
some notational complications to describe F-bases of the components Li of
M. Suppose that the i-th block Di is an n(i) x n(i) matrix, where n(i) > 1
is an integer. Since D is an s x s matrix, we have
s = n(l) H \-n(k).
L\ =Fei H hFe„(i),
L>2 = JFe n (i) + i + 1- Fe n (i) + n ( 2 ),
Li = F e n ( 1 ) + . . . + n ( j _ i ) + i + ■ • ■ + i ? e n ( 1 ) + . . . + n ( i _ i ) + n ( i ) ,
for v = 1 , . . . ,n{i).
It follows that M = Li © ■ • ■ © Lfc as an F[X]-module, as desired.
and
r- (pi,.-.,Pit) = (rpi,...,rp f c ),
where Pi,pi G P i , . . . ,Pfc,p'fc £ Pfc and r G R. An easy but long-winded
verification of the axioms (3.1) confirms that Pi x • • ■ x Pfc is an R-module.
7.7. Switching between internal & external 115
M = Lx + --- + Lk.
7.7.1 Proposition
The following assertions hold.
(i) If a left R-module M can be expressed as an internal direct sum
M = L\ © • ■ ■ ©Lfc, then M is isomorphic to the external direct sum
Li x ■■• x Lk.
(ii) If P = P\ x ■ • • x Pk is an external direct sum of left R-modules,
then P is an internal direct sum P = L\ © • ■ • © Lk of submodules
L\,..., Lk with Li = Pi for i = 1 , . . . ,k.
□
7.8 The Chinese Remainder Theorem
Our aim now is to show how a classical result from number theory form,
namely, the Chinese Remainder Theorem, leads to direct sum decomposi
tions of cyclic modules over Euclidean domains.
In its most familiar form, the theorem reads as follows. Suppose we
are given a pair of coprime positive integers m and n, and an arbitrary
pair of integers y, z. Then there is an integer x which satisfies both the
congruences
x = y mod m and x = z mod n.
We reformulate this assertion in the language of rings and modules.
First, notice that the congruence x = y mod z is equivalent to the equality
x = y in the cyclic Z-module Z m (see 1.11), and similarly x = z mod n
means that x = z in ^© 0 n(^feft^l?^f^^-/| n e a n i n S °f t n e notation "x"
7.8. The Chinese Remainder Theorem 117
Remark. At this point, the reader might expect to find a description of the
corresponding internal direct sum decomposition of R/Rbc. However, this
description requires some machinery that we are going to develop in a more
general setting in the next chapter, so we postpone a statement until more
tools are at our disposal - see Corollary 8.3.2.
7.8.2 Corollary
Let M = R/Ra be a cyclic module over a Euclidean domain R. Then M
is indecomposable if and only if a = upn, where p is an irreducible element
of R, u is a unit of R and n > 1.
7.4 Let D be the ring of 2 x 2 diagonal matrices over a field F, and let
e = e n and / = e 22 . Using Exercise 1.7, show that D = De © £>/ as
a left P-module.
Prove that the only D-module homomorphism from De to Df is
the zero homomorphism and hence that the /^-modules De and Df
are not isomorphic. (This contrasts to part (e) of Exercise 7.8.)
F.
Generalize these results to the ring o f n x n diagonal matrices over s
7.5 Let Pi x Pi be an external direct sum of P-modules. Define LJ : Pi x
P2 —» P 2 x Pi by u){pi,p2) = (p2,Pi)- Show that UJ is an isomorphism
of P-modules.
Given a set of modules { P i , . . . , Pk), k > 2 and any permutation
a of the integers 1 , . . . , k, prove that
Pi x ■•• x P f c ^ P Q ( i ) x ••• xP Q ( f c ) .
7.6 This exercise and the next anticipate some results that are developed
further in section 14.2. Let R be a ring and let M be a left P-module
with M = L © TV.
(a) Show that the canonical homomorphism IT : M —> M/L induces an
isomorphism r : N —> M/L.
(b) Let L = inc : N -> M be the inclusion map and let
Verify that these are all homomorphisms, and that the relations
and
£l7Ti + C27T2 = idp
hold.
Conversely, suppose we are given a collection of modules P, Pi, P2
and homomorphisms as above. Show that P = P\ x P%.
Generalize this exercise from 2 to k terms.
s 7-8 Let R be the ring of all n x n matrices over a field F. For each pair
of integers i, j = 1 , . . . ,n, let Cy be the matrix with entry 1 in the
(i,j)-th place and all other entries 0. The set of all such matrices is
sometimes called a set of standard matrix units for R (despite the fact
that the matrices e^ are not units of R).
(a) Show that the set of standard matrix units is a basis for R as a
vector space over F.
(b) Prove that
_ j ehk if i = j
ehi.eik - j 0 ii i^j ■
(c) For each i, j , let Ij = Reij. Deduce that Ij is the set of all matrices
A = ( 0 , . . . , 0, a,-, 0 , . . . , 0), where the j - t h column Oj is an arbitrary
vector in the column space Fn and all other columns of A are zero
vectors. (Thus Ij does not depend on the value of i).
(d) Show that R = I a left i?-module.
(e) For each pair of suffices j , k, define 6jk : Ij -> Ik by 9jk(x) = xejk-
Verify that 6jk is an isomorphism of left i?-modules.
(f) Show that R has no two-sided ideals apart from 0 and itself. (This
result generalizes Exercise 1.9.)
7.9 Direct products of rings.
s Exercise 7.8 can be generalized by introducing the direct product
of a set of rings. This is the construction for rings that corresponds
to the direct sum for modules. Let Ri,..., R^ be a set of rings, and
let R = Ri x • ■ ■ x Rk be the external direct sum of Ri,..., Rk as
additive groups. Define the product by
( r i , . . . , r f c ) ( s 1 , . . . , s f c ) = (risi,...,r f c s f c )
C 7.11 Let R be the set of all infinite matrices A = (aij) over a field F,
with rows and columns indexed by the positive integers, subject to
the condition that each row and column of A has only a finite number
of nonzero entries. Verify that R is a ring under the expected rules of
addition and multiplication. Find a set of left ideals I\s... so that R
is an internal direct sum 0 Ii as a left .R-module.
Let rnR be the set of all matrices A in R that have only a finite
number of nonzero entries. Show that mR is a two-sided ideal of R.
(This provides a contrast to Exercise 7.8 above.)
Chapter 8
123
124 Chapter 8. Torsion and the Primary Decomposition
Ann(m) = {r £ R | rm = 0} C R.
Ann(m) ^ 0,
T(M) = {m € M | m is torsion},
8.1.1 Proposition
(i) T(M) is a submodule of M.
(ii) M/T(M) is torsion-free.
to be 0 - see Exercise 8.6 below. However, the next result shows that this
does not happen in the cases of most interest to us.
8.2.1 Proposition
Suppose that M is a finitely generated R-module. Then M is a torsion
R-module if and only if Ann(M) ^ 0.
8.2.2 Corollary
Let F be a field, let A be an n x n matrix over F and suppose that M
is the F[X]-module obtained from Fn with X acting as A.
Then M is a torsion F[X)-module.
Proof Since the vector space of all n x n matrices over F has dimension n 2 ,
the n 2 + 1 powers I, A,..., An must be linearly dependent. Thus there is
a nonzero polynomial g(X) of degree at most n 2 with g(A) = 0. It follows
that g(X) is a nonzero element of Ann(M). □
8.3.1 L e m m a
Suppose that R is a Euclidean domain and that M is an R-module with
Ann(M) = Ra, a ^ O . Suppose also that a = be for coprime elements b,c
in R. Then the following hold.
(i) M = bM ® cM, where
m = 1 ■ m = b{xm) + c(ym),
and so
M = bM + cM.
If m e bM D cM, we have m = bl = en for some l,n € M and hence
cm = cbl = al = 0 and similarly bm = 0. Expanding 1 • m again, it follows
that m = 0, which gives bM D cM = 0 and hence M = bM © cM.
(ii) Suppose that x G Ann(6M). Then xbm = 0 for all m in M, and so
xb £ Ann(M) = Rbc. Since R is a domain, this means that x € i?c, which
shows that Ann(WW) C Re. But clearly Rc(bM) = 0, so that Ann(fcM) =
Re. The other equality follows by symmetry. □
Combining the above result with Theorem 6.5.2, we obtain the "inter
nal" version of the Chinese Remainder Theorem (7.8.1) that was promised
in the preceding chapter.
128 Chapter 8. Torsion and the Primary Decomposition
8.3.2 Corollary
Let M = R/Rbc be a cyclic R-module where b and c are coprime ele
ments in R. Then
M = bM®cM
with
bM * R/Rc and cM £* R/Rb.
□
8.4.1 Theorem
Suppose that R is a Euclidean domain and that M is a finitely generated
torsion R-module with Ann(M) = Ra, a ^ 0. Let p be an irreducible
element of R and write a = ppn with p coprime to p.
(i) If p is not a factor of a (so that a = p), then TP(M) = 0.
(ii) In general, the p-primary component Tp(M) of M is pM.
(Hi) There is a direct sum decomposition
M = pM®pnM.
8.4. The p-primary component 129
Proof
(i) Since p does not divide a, the elements a and pn are coprime in R
for any n > 1. Thus 1 = xa + yp n for x,y e R, and m = m • 1 = 0 for m in
T„(M).
(ii) By Lemma 8.3.1, we have M = pM ® pnM, and Ann(pM) = Rpn.
If m i , . . . , mt is a finite set of generators for M, then pm\,... ,pmt is a
finite set of generators for pM, which is thus a p-primary module according
to our definition. This shows that pM C TP(M).
To prove the reverse inclusion, suppose that m is an element of TP(M).
Then m = x + y with x £ pM and y e pnM. But y must be annihilated
by p n , since m is, and by p, since pnM is. Thus y = 0 and so m is in pM,
which gives the desired equality,
(iii) The decomposition is now immediate from Lemma 8.3.1. □
We can now give the complete primary decomposition of a torsion mod
ule.
8.4.2 Theorem
Suppose that R is a Euclidean domain and that M is a finitely generated
torsion R-module with Ann(M) = Ra, aj^O.
Let a = up™ .. .p2 be a standard factorization of a where u is a
unit of R and Pi, ■ ■ ■ ,Pk are distinct irreducible elements of R. Choose pt
so that a = p™ Pi for i = 1 , . . . , k.
Then, for each i = 1 , . . . ,k, the pi-primary component TPi(M) of M is
PiM, and M has the direct sum decomposition
M = TPl{M)®---®TPk{M).
Since Theorem 8.4.1 gives the equalities ptM = TPi(M) for all i, the result
follows. □
and so, using Proposition 7.7.1, we can express R/Ra as an external direct
sum of primary modules
Note that, by Theorem 7.3.1, the modules R/Rp*^ are all indecom
posable, so we cannot split R/Ra into smaller components, at least, not by
this technique. A uniqueness theorem to be proved later shows that any
other method of decomposing R/Ra into indecomposable modules must
give essentially the same result.
Exercises
8.1 (In these exercises, R is a Euclidean domain and all modules are
finitely generated R-modules, unless otherwise stated.)
Suppose that L is a submodule of M. Show that T(L) is a sub-
module of T(M) and that TP(L) is a submodule of TP(M) for any
irreducible element p of R.
More generally, if 9 : L —y M is an R-module homomorphism,
show that 9 induces a homomorphism T(9) from T(L) to T(M), and
likewise for Tp.
If 9 is a surjection, does it follow that T(9) is a surjection?
8.2 Suppose there is an isomorphism 9 : L —y M. Show that T(9) :
T(L) ->■ T(M) and Tp(9) : TP(L) -> TP{M) are also isomorphisms.
8.3 Show that every free .R-module is torsion-free.
8.4 Suppose that M = L® N. Prove that
(a) T{M) = T{L)®T(N)-
(b) TP(M) = TP(L) 0 TP(N) for any irreducible element p of R.
Generalize these results to the case that M has k > 2 components.
8.5 Suppose that P = L x N. Prove that
(a) T(P) 3 T(L) x T(N);
(b) Tp(P) =* TP(L) x TP(N) for any irreducible element p of R.
Generalize these results to external direct sums with k > 2 compo
nents.
8.6 Let the Z-module M = © i > 2 ^» be the infinite direct sum of the finite
cyclic modules Zj (Exercise 7.10).
Verify that every element of M is torsion, but that Ann(M) = 0.
132 Chapter 8. Torsion and the Primary Decomposition
8.7 Suppose that b and c are nonzero elements of R which are not coprime.
In this example, we outline an "elementary" argument which shows
that the direct product R/Rb x R/Rc is not cyclic. This should be
contrasted with the Chinese Remainder Theorem 7.8.1.
The fact that R/Rb x R/Rc is not cyclic can also be deduced
from the general uniqueness results which we shall obtain in Theorem
12.8.1 and section 12.9.
The argument is by contradiction, using the fact that a module
M is cyclic if and only if there is a surjection from R to M (section
4.11).
Verify the following statements, in which M = R/Rb x R/Rc.
(a) There is an irreducible element p of R which divides both b and c.
(b) There are positive integers m and n with TP(R/Rb) = R/Rpm,
TP{R/Rc) =* R/Rpn, and
Presentations
Our results to date have given us a reasonable hold on the theory of cyclic
modules over a Euclidean domain R. Combining Theorem 6.4.1 with the
factorization theorems for elements of R, we know that a cyclic .R-module
has the form R/Ra for an element a of R which is unique up to multipli
cation by a unit, and we can calculate the primary decomposition of R/Ra
as in section 8.5. The problem we now face is to extend these results to
arbitrary finitely generated i?-modules. To provide some motivation for
what follows, we take another look at how we recognize a cyclic module as
being isomorphic to one of the form R/Ra.
The statement that a module M is cyclic tells us that it has a single
generator, say m; then M = Rm. The generator will satisfy some "rela
tions" , that is equations of the form rm = 0 for r £ R. If the only such
equation is the trivial one, Om = 0, then {771} is a basis for M, which means
that M is free and isomorphic to R.
If there are nontrivial relations, then there is a "fundamental" relation
am = 0 with the property that all other relations are consequences of this
fundamental relation, that is, they take the form (xa)m = 0 for x 6 R. This
assertion is simply a reinterpretation of the fact that the set of coefficients r
with rm = 0 forms the annihilator ideal Ann(m) of m, which is a principal
ideal Ra. The First Isomorphism Theorem now assures us that M = R/Ra.
It is the analysis of generators and relations that provides the key to
our description of /^-modules in general. Suppose that M has a finite set
{ m i , . . . , mt} of generators, which means that each element m in M can be
written
m = rivfix ~\ (- rtmt for some r j , . . . , rt e R.
Then there are usually some relations between the generators, that is, iden-
133
134 Chapter 9. Presentations
x = riei H \-rtet
for some unique members r\,..., rt of R (section 5.1). Thus, given a pre
sentation 0 of a module M, each element m of M has the form
m = 6{x)
= ri0(ei) + --. + r t 0(e t )
and then
0(ei) = m i , . . . , % ) =mt.
Notice that a module has many different presentations, since each choice of
a set of generators gives rise to a presentation.
9.2. Relations 135
Examples.
(i) A cyclic module R/Ra has a presentation TC : R -> /?/i?a, 7r(r) = f = r - I
and also a presentation —?r: r t-> r • (—1).
(ii) A module will have generating sets of different sizes, and so it will have
presentations involving free modules of different ranks. For example, take
M = R/Ra x R/Rb, the external direct product of two cyclic modules,
and put mi = (T, 0) and m 2 = (0,T). Then M = Rmi ® Rm2, and there
is a presentation p : R2 -» M given by
9.2 Relations
Suppose that 9 : Rl —> M is a presentation of M, and write
0 = 7nmi + h 7«mt \
: \ . (9.2)
0 = 7i s mi H h 7tsmt J
We therefore say that the set of relations {pi,..., ps} is a set of defining
relations for M.
For example, the cyclic module R/Ra has one defining relation, p\ =
aei, while the direct sum R/Ra x R/Rb has two defining relations, pi = ae\
and p2 = be%-
K = Rpi + ■ ■ ■ + Rps
K = 3Zx5ZcZxZ
M =■ Z 3 0 = Z 2 x Z 3 x Z 5
by using the Chinese Remainder Theorem together with the results in sec
tion 8.5.
The next result tell us that a set of defining relations can always be
taken to be finite.
9.4.1 Theorem
Let R be a Euclidean domain and let K be an R-submodule of a standard
free R-module Rf of rank3e>[jphejtot8cliWatm#tee, of rank s with s <t.
138 Chapter 9. Presentations
x - bw € Ker(tr) n K.
x = (x — bw) + bw £ L + Rw,
which gives
K = L + Rw.
If x £ L n _Rw, then x = bw with ba = 0. Thus 6 = 0 and so x = 0. This
establishes that
K = L © to.
Now L is a submodule of Ri~1, so the induction hypothesis tells us that
L is free of rank h say, with h < t — 1. Let { c j , . . . , c^} be a basis of L.
Since .Rw has basis the single element w (and has rank 1) , K is free with
basis { c i , . . . ,Ch,w} and rank s = h + 1 < i. D
Remarks.
(i) This theorem assures us that a finitely generated R-module with t gen
erators has a presentation that requires s < t defining relations,
(ii) In vector space theory, a generating set can be reduced to a basis by
omitting elements. Such easy arguments cannot be expected to work for
modules over Euclidean (&wpfyfl}§h1<mfcM$eftk& generating set {2,3} of Z.
9.5. The presentation matrix 139
(iii) The argument in the proof of the theorem is, in principal, constructive.
Given an explicit set of generators
{Pl,---,Pu}
9.4.2 T h e o r e m
Let R be a Euclidean domain and let M be a finitely generated R-module.
Then every R-submodule of M is also finitely generated as an R-module.
Pi = 7 n e i H H 7ue, + h 7ne*,
Pi = 7 i j e i + • • ■ + jijei H + jtjet,
Ps = 1Q9PW&tafl'M0p4al- ■ + itset
140 Chapter 9. Presentations
/ 7n • • m ■ ■■ l u \
7« • lij lis
notice that the coordinate vector of pj gives the j'-th column of the matrix.
Despite Theorem 9.4.1, we allow the possibility that s > t. The reason for
this is that a module may be presented to us with surplus relations, but it
may be far from obvious which relations can be discarded.
We can reverse our point of view, obtaining a module from a matrix.
Given any t x s matrix T over R, we define the submodule K of Rf to
be that with generators p\,... ,ps as above, and then define M to be the
quotient module M = Rt/K. By construction, T is a presentation matrix
for M.
7(^>i) = Pi = 7 n e i H h 7«et,
(9.3)
l{bs) = Ps = 71*ei H + ItsGt-
9.6.1 T h e o r e m
The presentation homomorphism 7 is injective if and only if the set
{pi,..., ps} of defining relations forms a basis of the relation module K =
Ker(0).
9.7. F[X]-module presentations 141
Proof We know that the defining relations generate K, so they form a basis
precisely when they are linearly independent. But pi = -y(bi) for all i by
definition of 7, and since b\,..., bs is the standard basis of Rs, the defining
relations are linearly independent if and only if 7 is injective. □
7T : F[X]n -> Fn
X -a j] (9.4)
r= -Q.il —a;n
\ ~0.nl L
nj /
which is the characteristic matrix of the matrix A.
Let XI be the product of X with the n x n identity matrix 7, that is,
the n x n diagonal matrix with diagonal entries all equal to X. Then we
can write the characteristic matrix more succinctly as
XI- A. (9.5)
The determinant
det(r) = det{XI - A)
is the characteristic polynomial of A, which plays an important role both
in elementary linear algebra and in our analysis of the structure of F[.X']-
modules. A straightforward expansion shows that the characteristic poly
nomial is monic, with
det(XI - A) = Xn - (on + ■ • • + ann)Xn-1 + ■■■ + ( - 1 ) " det(,4).
9.8 F u r t h e r developments
The attempt to describe modules in terms of presentations leads to some
of the most fruitful areas of mathematics, in that it provides a method of
characterizing various types of rings and modules.
The method can be illustrated by the key result of the chapter, Theorem
9.4.1, which assures us that any finitely generated module M over a Eu
clidean domain R has a free relation module K, and that the rank of K is
at most the number of generators of M. This result holds for commutative
principal ideal domains ([Cohn 1], §10.5) and, in fact, it provides a charac
terization of such rings. Suppose that the theorem holds for a commutative
ring R. Then a nonzero ideal I oi R must be a free .R-module with one
generator and so I — Ra is principal. Moreover a cannot be a zero divisor
in R, since the obvious surjection from R to 7 cannot have a nonzero free
kernel. Thus R must be a principal ideal domain.
If the ring R is not a principal ideal domain, then the .R-modules that
do have a presentation with a free relation module form an interesting type
of .R-module.
Exercises 143
Exercises
9.1 Let R be a commutative domain and suppose that the R-module M
has a presentation 7r : R* —> M with square presentation matrix T
and that det(r) ^ 0. Using the formulas given in Det 8 of section
5.12, show that the relation module Ker(-7r) for M contains
det(r)ei,...,det(r)et.
Diagonalizing and
Inverting Matrices
145
146 Chapter 10. Diagonalizing and Inverting Matrices
Recall that the entries of T are given by the coefficients 7^ in the defining
relations of Eq. (9.2):
0 = 7 n m i + ••• + 7 u m t ,
(10.1)
0 = 7i s mi + ■•■ + 7 t s m t .
The effect of column operations is transparent. Each column of T is
made up of the coefficients from an individual relation, so an elementary
column operation on T corresponds to the "same" operation on the rela
tions. Here is a detailed list.
exchanging columns i and j :
exchanging relations i and j
(3) Suppose that we multiply row i by the unit u of R. The new j-th
relation is
Using column operations (and omitting some steps), we make the transfor
mations
r<ll)All)
giving the relations
0 = mi + 3m,2
0= 2m 2 .
Subtracting 3 times the first row from the second gives the matrix
(i!)
and relations
0 = m'1
0= 2m'2
with
m1 = m,\ — 3m2, m'2 = ra2-
Thus the Z-module defined by the relations (10.3) is Z 2 .
10.3. A matrix interpretation 149
10.3.1 Lemma
(i) PT = (pI)T.
(li) TX = Y{Ix).
(Hi) Let p~l and x _ 1 be the elementary operations which reverse the
effects of p and x respectively. Then the matrices pi and Ix are
both invertible, with inverses p~xI and Ix-1 respectively.
(iv) (pT)x = p ( r x ) .
Proof
(i) The proof is by direct calculation. Since only two rows of I and T
are alterd by the row operation, it is enough to make the verification when
r is 2 x s and / is 2 x 2. (The doubtful reader can fill in the unchanging
rows for reassurance.) Suppose, for example, that p is of type 2, "add r
times row i to j , i ^ j " . Ignoring the unaffected rows (and assuming i < j),
we have
1 0
pi
( r 1 )
and
7»i 7i2 ••• liS
(pi)r ) ■
(
which is obviously pT. The calculations for the other types of elementary
row operation are easier.
(ii) The argument for column operations is similar, working with t x 2
matrices.
150 Chapter 10. Diagonalizing and Inverting Matrices
(iii) We have
{p-ll){pl) = p~\pl)
= {P~1P)I
= I
and similarly (pl)(p~1l) = I, confirming that (pl)~l = p~lI-
The calculation for x is much the same, working on the right rather
than the left.
(iv) In matrix terms, we require (pI)(T ■ Ix) = (pi ■ r ) ( J x ) , which is
true since matrix multiplication is associative. □
P = Ph ■ ■ -Pi,
and
pr = ph(...(Plr)...).
Notice that the elementary operations are listed from the right to the left
in the product since row operations operate on the left of matrices.
Likewise, a column operation x on T is the product of a sequence
Xii • • • tXh of elementary column operations, again performed in the given
order. We can write x as
X = Xi---Xk
where the terms are ordered from left to right, since column operations act
on the right of matrices.
The ordering of the elementary row operations is very important, since
the same operations in a different order will give a different row operation.
For example, let p\ be "exchange row 1 and row 2", and let pi be "add row
1 to row 2". Then
P2PJ= ( J J)
while
10.4. Row & column operations in general 151
10.4.1 Lemma
(i) Let p = Ph • ■ ■ P\ be a row operation, where each pi is an elementary
row operation. Then p is invertible, with inverse
p^=p-^...p-h\
pI = (phI)...(PlI).
(Hi) The matrix pi is invertible over R, with inverse (p^ I)... (p~^ I),
(iv) Let x — Xi ■ ■ ■ Xk be a column operation, where each \i *s an e^e~
mentary column operation. Then x *s invertible, with inverse
X" 1 =Xk1---Xil-
(v) There is a matrix equation
IX = (IXi)---(IXk).
Proof
(i) Multiplying pn ■ ■ ■ P\ by p\l ... p^1, on either side, we see that all
the terms cancel, so we are left with the identity row operation, which does
nothing to any matrix that it operates on. So the product of the inverses
is indeed p~l.
(ii) This follows by induction on h from part (i) of Lemma 10.3.1. We
have
pi = (ph{Ph-i ■ ■ ■ Pi))I
= {PhI){{Ph-\ ■ ■ ■ P\)I)
GQp0gmfrM3terial. (PlI)).
152 Chapter 10. Diagonalizing and Inverting Matrices
(iii) Immediate from the above, using the fact that, for invertible ma
trices A,B, we have (AB) _ 1 = B _ M _ 1 .
The assertions about column operations have similar proofs, and the
final claim follows from the associativity of matrix multiplication. □
5l 0 ■ 0 0 ■• ■
1 52 ■ ■
° \
0 0 0 ■■ • 0
A = 0 0 • 5T 0 ■• • 0
0 0 • ■ 0 0 ■■ • 0
VV oo o0 • • 0 0 •• • o/
whose nonzero entries 5i,... ,5r satisfy the relations
A = diag(<5i,... A , 0 , . . . , 0 )
when convenient.
Now let T be any s x t matrix over a Euclidean domain R. An in
variant factor form or Smith normal form for T is any diagonal matrix
A = diag(<Ji ,...,Sr,0,...,0) that is itself in invariant factor form and which
is related to T by an equation
PTQ = A
The invariant factors are not unique, since any of them can be multiplied
by any unit of R. However, this is the only type of change permitted. We
will prove this assertion in the next chapter, Theorem 11.3.1, together the
uniqueness of the rank. Our immediate task is to show that any matrix
does have an invariant factor form.
Proof
We start by observing that the second assertion is a consequence of
the first. Suppose that pT\ = A where p is a product of elementary row
operations and x is a product of elementary column operations, and put
P = pi and Q = I\ for suitable identity matrices. Then, by Lemma 10.4.1,
both P and Q are invertible over R and PTQ = A.
The proof of the first assertion is by double induction. We induce on
both the size of the matrix and the size of its minimum nonzero entry. We
present the argument in a series of steps; to avoid over-elaborate notation,
we behave as programmers and allow the meaning of T and some other
matrices to vary during the argument.
The fact that we use operations only of types 1 and 2 will be clear from
the method. There are two starting cases in which we need not do anything.
If T = 0, then it is in the desired form with r = 0, and if T is 1 x 1 it is
obviously diagonal.
We define the "size" $(T) of the minimum entry by using the function
if : R -> Z (section 2.1). Suppose that T = (7^) is nonzero. Then we put
* ( r ) = m i n t e d ) I 7 i ^ 0}.
no need to exchange rows, that is, we perform the identity row operation,
and likewise if k = 1.
Step 2. We can now assume that 3>(r) = <£>(7n). If s = 1, that is, T
has only one column, go straight to step 3. If not, proceed as follows.
For j = 2 , . . . , s, use the division algorithm to write
and then let Xj be the column operation "subtract q\j times column 1 from
column j " for j = 2,..., s.
If some 7£ ■ ^ 0, then we have a new matrix T' with 3>(r') < 3>(r). By
induction hypothesis, we can reduce T' and hence T to the desired diagonal
form.
If j[- = 0 for all j , we turn to row operations.
Step 3. If there is only one row, the preceding arguments will diagonalize
T, so we may assume that t > 2. For i = 2 , . . . , t, write
/ 711 0 •■■ 0 \
0
r
= : r
V 0 )
Step 4- If by chance T' = 0, we are finished. If not, we perform long di
visions to see whether or not 711 divides the entries 7^ of the submatrix V.
It will be convenient to index the entries of T' according to their positions
in the larger matrix T rather than their positions in V itself.
For i,j > 2, write
After a finite number of trips around this loop, reducing $ ( r ) each trip,
we must arrive at the stage where
/ 711 0 ... 0 \
0
r= • r
'
/ 12 9 12 \
T= 3 2 4 .
CopyrtgHtid MaffiieJ
10.7. A computational technique 157
12 9 12 1 0 0
3 2 4 0 1 0
r/ / 12 8 22 0 0 1
1 0 0
0 1 0
0 0 1
3 2 4 0 1 0
12 9 12 1 0 0
12 8 22 0 0 1
1 0 0
0 1 0
0 0 1
2 3 4 0 1 0
9 12 12 1 0 0
8 12 22 0 0 1
0 1 0
1 0 0
0 0 1
2 1 0 0 1 0
9 3 -6 1 0 0
8 4 6 0 0 1
0 1 0
1 -1 -2
0 0 1
1 2 0 0 1 0
3 9 -6 1 0 0
4 8 6 0 0 1
1 0 0
-1 1 -2
0 0 1
1 0 0 0 1 0
3 3 - 6 1 0 0
4 0 6 0 0 1
"*■ 1 - 2 0
-1 3 -2
Copygightqfi Material
158 Chapter 10. Diagonalizing and Inverting Matrices
1 0 0 0 1 0
0 3 - 6 1 - 3 0
0 0 6 0 - 4 1
1 -2 0
-1 3 -2
0 0 1
1 0 0 0 10
0 3 0 1 - 3 0
0 0 6 0 - 4 1
1 -2 -4
- 1 3 4
0 0 1
Thus the invariant factor form of T is
1 0
°\
A = 0 3
\ 0 0 6/
° '
the row operations are recorded as
0 1 0
P = pl = 1 -3 0 ,
Vo -4 li
and the column operations as
/' l -2 -4\
Q = ix = -l 3 4
\ ,° 0 1/
10.8.1 Lemma
Let A = diag(<5i,..., 6r, 0 , . . . ,0) be atxs diagonal matrix with nonzero
entries 6\,... ,8r contained in a Euclidean domain R.
Then the following statements are equivalent.
(i) A is invertible ovec^yrighted Material
10.8. Invertible matrices 159
Proof Suppose that A is invertible. If, say, t > r, then A has a row of
zeroes, and so does A • A - 1 = /, a contradiction. Thus t = r and likewise
s = r. The fact that all the terms Si are units now follows by direct
calculation, or alternatively, from the fact that det(A) = <5X . ■ -Sr is a unit
in R (Theorem 5.12.1).
The converse is obvious. □
This lemma gives several characterizations of invertible matrices which
we list in the next theorem.
10.8.2 Theorem
Let r be a t x s matrix with entries in a Euclidean domain R.
Then the following statements are equivalent.
(i) r is invertible over R.
(ii) T is a square matrix and the invariant factor form ofT can be taken
to be the identity matrix.
(Hi) T — pi for some row operation p.
(iv) V = I\ for some column operation x-
Proof (i) => (ii): By Theorem 10.5.1, there are invertible matrices P, Q so
that the invariant factor form of T is A = PTQ. Since T is invertible, so
is A. Thus A and hence T are square matrices. Also, 6i,...,fir are units
in R, and so, by multiplying column i by S~ for each i, we can take the
invariant factor form to be the identity matrix /.
(ii) => (iii): We have PTQ = I, where P and Q are invertible square
matrices, say r x r. Thus T = P~1Q~l.
By construction, P is obtained from the identity matrix / by applying
a sequence of elementary row operations, and P _ 1 is obtained by applying
their inverses - see Lemma 10.4.1. Thus P~l = p'l for some row operation
P'-
By the same lemma, Q l = (Ixi) ■ ■ ■ (IXk) f° r some sequence of elemen
tary column operations X\i■ ■ ■ iXk- But for each such elementary column
operation, there is an elementary row operation pi so that pil = Ixi - this
is easily seen by evaluating Ixi for each of the three types of elementary
column operation. Thus, writing p" = pi... pk, we have T = p'p"I.
(iii) => (iv): Arguing as in the previous paragraph, we can replace each
elementary row operation in p by an elementary column operation, and so
find a column operation x with pi = I\-
(iv) => (i): Already prove^cipj^Jgfttedl AfefeA'a/ □
160 Chapter 10. Diagonalizing and Inverting Matrices
from which
- for some principal ideal domains there are invertible matrices which can
not be reduced to an identity matrix by any sequence of row and column
operations. However, examples of this phenomenon are surprisingly hard
to find; some are given in [Grayson] and [Ischebeck].
The diagonalization results also extend to noncommutative Euclidean
domains ([B & K: IRM], §3.3) and to noncommutative principal ideal do
mains ([Cohn: FRTR], Chapter 8).
When the coefficient ring is not a principal ideal domain, we cannot
expect that a matrix will have a standard form that is as simple as the
invariant factor form. Even if the ring of scalars is close to being a Euclidean
domain, the analogue of the invariant factor form can be rather complicated,
as can be seen from [L & S].
Exercises
10.1 Let M be a Z-module with generators mi,m2 and relations
m\ + 2m,2 = 0,
2mi + 3m 2 = 0.
2mi + 2m 2 + 3m 3 = 0,
4mi + 4m 2 + Im-j, = 0,
10mi + llm2 = 0.
mfiP&R&&& 4fefcfl#i + m 2 = 0.
162 Chapter 10. Diagonalizing and Inverting Matrices
Fitting Ideals
/ 3 2 1 0 \
T= 10 5 - 5 15 I ,
\ -20 -10 10 -30 /
CopyrightdSft/laterial
164 Chapter 11. Fitting Ideals
11.1.1 L e m m a
Let A = diag(<5i,..., 5r, 0 , . . . , 0) be a matrix in invariant factor form.
Then the Fitting ideals of A are
where
1 < i(l) < . . . < i(h) <r
is an ascending sequence of integers. The divisibility condition on the en
tries of A now shows that 8\... 5h divides any such product and so must
generate Fit^(A).
Finally, if h > r, then any hxh submatrix contains a row of zeroes and
so has zero determinant. Copyrighted Material □
11.2. Elementary properties 165
11.2.1 L e m m a
Let r be atx s matrix, let p be an elementary row operation on V and let
X be an elementary column operation on T. Then for h = 1 , . . . , min{i, s},
Fith(/oT)=Fitfc(rx)=Fith(r).
Proof We give the argument only for a row operation p. Consider the effect
of p on the determinant of an h x h submatrix E of T. If p does not involve
any rows of E, nothing happens to the determinant. If p involves two rows
of E, then p is effectively a row operation on E and, by standard properties
of determinants (section 5.12), it changes the determinant as follows.
• If p exchanges two rows,
det(pE) = - d e t ( E ) .
det(pE) = det(E).
det(pE) = udet(E).
Finally, we have the situation when p involves two rows, one of which,
say row i, contributes to E and the other, row j , does not. There are now
two cases in which the determinant might be changed.
• Suppose p exchanges rows i and j . Then pE is an h x h matrix,
formed from the same columns as E, in which the entries of E from
row i of T are replaced by the corresponding entries from row j of
T. Thus there is an h x h submatrix E' of T so that pE and E' have
the same rows, but perhaps in a different order, which gives
det(pE) = ±det(E').
Since the Fitting ideal Fit h(pT) of order h is generated by the determi
nants of all the h x h submatrices of T, we can now see that Fith{pT) C
Fith(r). But p is invertible, so we also have Fit/»(r) C Fith{p£)- D
11.2.2 Corollary
Let r be at x s matrix.
(i) Given a product p of elementary row operations on T and a product
X of elementary column operations onT, we have
F i t h ( p r x ) = Fith(r)
Proof (i) This is immediate from the Lemma together with part (iv)
of Lemma 10.3.1.
(ii) By Theorem 10.8.2, we can write P = pi and Q = I\, so the claim
follows. □
11.3.1 Theorem
Let r be a matrix over a Euclidean domain R and suppose that
A = diag(<5i,... A , 0 , . . . , 0 )
and
A' = diag(*i,...,£.,0,... > 0)
are both invariant factor forms of T.
Then r = r', and there are units U\,..., ur of R with
5i — « i 5 i , . . . , S'r = ur5r.
11.4. The characteristic polynomial 167
Proof By Theorem 10.5.1, we have A = PTQ and A' = P'TQ' for invertible
matrices P,P',Q and Q', so that A' = P'P^AQ^Q'.
Thus A and A' have the same Fitting ideals by the preceding corollary,
which means that r = r' and that
for h = 1 , . . . ,r.
Taking h = 1 we get 5[ = ui<5i for some unit wi of the domain R, and
the rest follows easily. □
Standard choices. In the ring Z of integers, we take an invariant factor
to be positive, and in a polynomial ring F[X], we take an invariant factor
to be monic. With these choices, the invariant factor form of a matrix is
unique.
The preceding results yield the solution to the equivalence problem that
we posed in section 10.6.
11.3.2 Corollary
Let r and V be t x s matrices over a Euclidean domain R. Then the
following statements are equivalent.
(i) r and V are equivalent,
(ii) Fith(r) = Fit f t (r') forh = l,.. .,min(t,s).
(Hi) T and V have the same invariant factor forms (up to multiplication
of their entries by units).
□
11.4 The characteristic polynomial
Let A be an n x n matrix over a field F, and let M be the module over the
polynomial ring F[X] which is defined by A acting on the space Fn. By
section 9.7, the presentation matrix of M has the form V = XI — A, where
I is an n x n identity matrix. For matrices of this type, it is often more
convenient to compute the invariant factors through the Fitting ideals.
It is clear that the n-th Fitting ideal of XI — A is
ftpyfflte&fM&r^A). (n.2)
168 Chapter 11. Fitting Ideals
Since the invariant factors are themselves polynomials, we can see, just
by counting their degrees, that many of them are likely to be constant
polynomials. Indeed, suppose that no invariant factor is a constant. Then
all must have degree 1, and by the divisibility condition S\ | • • • | <5„, they
are all equal to X — a for a fixed constant o. Thus A = al.
At the other extreme, consider the rational canonical block matrix
(° 0
0
• •
• •
0
0
0
0
-/o
-/l
\
1
0 1 • • 0 0 -/a
c
0 0 • • 1 0 — fn-2
0 ■ • 0 1 -fn-1 )
51 = ■ ■ ■ = $n-i = 1 and 6n = f.
Exercises
11.1 Use the Fitting ideals to compute the invariant factors of the integer
matrix discussed in section 10.7:
12 9 12 \
3 2 4
- 8 22 /
V 12
11.2 Compute the Fitting ideals of the following matrices, and hence con
firm your calculations in the exercises to Chapter 9.
1 2 \
(a) 3
I2 )
2 4 10 \
/
2 4 11
(b)
7
\ 3
2 '\
oj
(c) ( 1
I2 1 J
3 2 \
/
(d) 3 6 ■
\2 / a
11.3 In the last part of the above exercise, show that the three cases a =
0,1, 2 cover all the rS@flB5i9fiel<3foVf^a^ariant factors of the matrix.
170 Chapter 11. Fitting Ideals
/ 1 0 o \
11.4 Let A = I 0 1 a ) be a matrix over a field F. Discuss the pos-
\ 0 0 b )
sibilities for the invariant factors of XI — A as a and b vary.
11.5 For those who relish a challenge!
Let M be a Z[i]-module with generators mi, m 2 , rnz and relations
T h e Decomposition of
Modules
Now that we have the invariant factor form of a matrix at our disposal,
we can exploit it to obtain a very nice presentation for a module M over
a Euclidean domain. This presentation leads to a decomposition of the
module as a direct sum of cyclic modules, which in turn allows us to find
the torsion submodule T{M) and the p-primary components TP(M) of M,
and to show that the quotient module M/T(M) is in fact a free module.
We also show that the descriptions which we obtain are unique, subject
to some conditions on the way in which we arrange the cyclic summands of
M.
As applications of our results, we obtain the structure theory of finitely
generated abelian groups and some properties of lattices in R n .
Throughout this chapter, R is a Euclidean domain and all .R-modules
are taken to be finitely generated.
We start with a description of the submodules of a free module.
171
172 Chapter 12. The Decomposition of Modules
Write
Pi = 7 n e i + 1" ltiet
(12.1)
Ps = 7i*ei + ■ ■ ■ + 7t»Ct
as in Eq. (9.1), where { e j , . . . , e t } is the s t a n d a r d basis of Rl, and p u t
/ 7n •• 7ia \
T= : •.. :
\ 7u ■•■ Its )
l{b\) = p i , . . . , 7 ( 6 s ) = ps
K = <?i A + 92S&2 + • • ■ + 9 SS 6 S
e
* ^Wti^m^Mitenaf P"e'f
12.1. Submodules of free modules 173
l
To find E' in terms of E, we must calculate the inverse P = (p^A Then
. P - 1 = P E , E ' by the results of section 5.6, and
e
l = P l l e l +P21 e 2H \-Ptlet,
4 = P l 2 l + P~22 2 + ••• +Pt2 e *>
e e
: (12-4)
e
t = Puei +P~2te2 H +P«et.
We can now obtain a very useful description of a submodule of a free
module.
12.1.1 Theorem
Let K be a submodule of the standard free R-module i?*. Then K is free,
and we can find a basis E' = { e ^ , . . . , e't} of Rl and elements 5\,..., Sr of
R so that
8\e\,... ,8Te'r
is a basis of K, with 5\\ ■ ■ ■ \ 5T.
1(b'1)=S1e[,...,1(b'r)=5re'r
is a set of generators for K. But this set must also be linearly independent,
since the set e\,..., e'r is already linearly independent. □
A computation. The techniques given in Chapter 10 enable us to carry
out numerical calculations. For example, suppose that K is the submodule
of 1? with three generators
pi = 12ei + 3e 2 + 12e3, ]
p2 = 9ei + 2e 2 + 8e 3 , > (12.5)
pz = 12ei + 4e 2 + 22e 3 . J
The corresponding matrix is
/ 12 9 12 \
T= 3 2 4
Copyr'fahlSd l&latffiall
174 Chapter 12. The Decomposition of Modules
and the calculations in section 10.7 show that T has invariant factor form
1 0 0
PTQ = A = | o 3 o
0 0 6
with
0 1 0 1 2 -4
P- 1 -3 0 and Q = -1 3 4
0 -4 1 0 0 1
We find that
3 1 0 \
>-i
1 0 0
V4 o i;
Thus the basis
= 3ei +e2 + 4e 3
= ei (12.6)
= e2
of 1? gives the basis
e
i> 3e 2 , 6e 3
of K.
and the relation module for M is the kernel Ker(#), which has generators
Pi = 7 n e i H +7tiet,
(12.8)
e
P^
bopfaffited ±3te t-
Material
12.2. Invariant factor presentations 175
Put
12mi + 3 m 2 + 12m 3 = 0, ]
9mi + 2 m 2 + 8 m 3 = 0, i (12.9)
12mi + 4m2 + 22m3 = 0. J
The calculations in the preceding section show that we can find new gen
erators for M, given by
Simi = 0 , . . . , Srmr = 0.
It may happen that the presentation starts with some terms whose co
efficients Si, i = 1 , . . . , h, are units in R, since a presentation matrix for M
may well have unit invariant factors. As a unit can always be replaced by
the identity element of R, the presentation begins
m i = • • ■ = nth = 0.
12.3.1 Theorem
Let R be a Euclidean domain and let M be a finitely generated R-module.
Then the following statements are true.
(i) M has a set of generators m\,... ,mt so that M is the internal direct
sum
M = Rmi © ■ • ■ © Rmr © Rmr+i © • • • © Rmt
with
Run & R/RSi for i = l , . . . , r
and
Rmj = R for j = r+ l,...,t,
where Si,..., Sr are non-units in R with Si \ • • ■ \ Sr.
(ii) The torsion submodule of M is
by
ar+i = ■ ■ ■ = at = 0
and
aai £ RSi for i = 1,... ,r.
Thus T(M) C Ami ffi - • •ffii?m r . However, the divisibility conditions on
5\,..., Sr show that
ST{Rmx © • • • © RmT) = 0,
The elements 5\,..., Sr are called the invariant factors of M and the
integer w is called the rank of M. The direct sum decomposition of M
into cyclic summands is known as the invariant factor decomposition of M.
The uniqueness of the invariant factors of M will be established in Corollary
12.8.2.
Notice that the uniqueness theorem for the invariant factors of a matrix
does not lead directly to the corresponding result for modules, the point
being that a given module can have many unrelated presentation matrices.
The invariant factor decomposition may also be less precisely referred
to as the cyclic decomposition, although it contains more information than
simply telling us that there is a cyclic decomposition of a module.
r=
1 3
10
2
5 -5
1
°\
15
\ -20 -10 10 -30 /
that we introduced in section 11.1. The Fitting ideals of this matrix are
Fiti(r) = z )
Fit 2 (r) = 5Z V (12.12)
12.5. The primary decomposition 179
A = diag(l,5,0).
Z 5 x Z,
so we have r = 1 and w = 1.
The next result, although also an illustration, is important enough to
be recorded separately.
12.4.1 Proposition
Let F be a field, let C be the rational canonical block matrix associated
to the polynomial f = f0 + f1X H h / n _ i X n _ 1 + Xn, and let M be the
module over the polynomial ring F[X] which is defined by the action of C
on Fn.
Then
M = F[X]/F[X]f(X).
(° 0 • • 0 0
0 • • 0 0
-/o
-/l
\
1
0 1 ■ • 0 0 -h
C
0 0 • ■ 1 0 — fn-2
0 ■ • 0 1 -fn-1 /
As in section 11.4, the invariant factors of T are
with
TPi (R/RS) « R/Rpfj) for j = 1 , . . . , k.
Now we return to the general case. Suppose that M = R/RSi © • ■ ■ ©
R/R5r is the invariant factor decomposition of M, and let
5T = u r p j ---Pk
be a standard factorization of 5r. Then for i < r we can write
r n(i,l) n(i,k)
Si =Uip^ ■■■Pk ■
0<n(l,j)<---<n(r,j).
M = Rg1%x © • ■ • © Rgr:k
with
RgtJ^R/Rpf-j) for all i,j.
Suppose that we have already found a set of generators m\,..., mr which
give the invariant factor decomposition of M:
so t h a t
n ( l , l ) = 0, n ( l , 2 ) = l, n ( 2 , l ) = l , n ( 2 , 2 ) = l .
T h e generators are then
3m
£?i,i = i > 9i,2 = " i i , 92,1 = 3 m 2 , 92,2 = 2 m 2
with relations
where the exponents are all nonzero and are listed in nondecreasing order
for each pj, t h a t is,
z(j,l)<---<zU,y(j)).
Essentially, the method is to construct t h e "largest" invariant factor by
taking the product of largest powers of each irreducible element t h a t occur
in t h e list, then to find t h e next largest in terms of the remaining powers,
and so on.
More formally, we proceed as follows. T h e number of terms in which
Pj occurs is y(j). Let t h e integer r be the maximum value of y(j) for
j = 1,... ,k, and put
n rr z kk
<( >!)
>*) l±^Htihlamtt&S= z=(k>y(
nh^lMriM^^) ^v())))
12.8. The uniqueness results 183
and
°r — Pi • ■ • Pk
Next, write
12.8.1 Theorem
Let M be a p-primary R-module. Then the following hold.
M ^ R/Rpn{1) x • ■ •x R/Rpn^
(ii) The set {pn^\ ■ ■ ■ ,pn^} is both the set of elementary divisors of M
and the set of invariant factors of M, and it is uniquely determined
by M.
n ( l ) = ■■■ = n(z) = 1.
If all the exponents n(i) are greater than 1, we take z = 0, and, in any case,
z <y.
For any exponent n > 1, we have Rp/Rpn = R/Rp71-1 (Exercise 6.4),
while p{R/Rp) = 0. Hence
n(z + l ) - l , . . . , n ( y ) - l
n(z + l ) , . . . , n(y)
(R/Rpn)/p(R/Rpn) a R/Rp,
z = y- (y- z),
12.8.2 Corollary
Let M be an R-module. Then the invariant factors of M are uniquely
determined by M, except that each of them can be multiplied by a unit of
R. ' ' " □
12.9 A summary
As our description of the structure of a module has emerged piecemeal,
here is a brief summary of the main results.
Let M be a finitely generated module over a Euclidean domain. We
start by decomposing M as a direct sum T(M)®F{M), where T(M) is the
torsion submodule of M and F{M) is a free complement, as in Theorem
12.3.1. The submodule T(M) is absolutely unique, as we can see from its
definition as a subset of M. Exercise 12.1 shows that the free complement
F(M) is not usually unique as a subset of M, but it does have a unique
rank.
The next step is to express T(M) in invariant factor form
as we have just seen, the invariant factors are unique apart from changes
of the form 5't = UiSi for units u, of R. The corresponding internal direct
sum decomposition of T(M) is usually far from unique, as can be seen by
considering all possible bases of the vector space R/Rp x R/Rp - any such
basis gives an internal direct sum decomposition of R/Rp x R/Rp as an
.R-module.
The torsion submodule can further be decomposed into p-primary com
ponents TP(M) for irreducible elements p of R. These components are
absolutely unique since they are specific subsets of M. There is a unique
finite set pi,...,pk of irreducibles for which TPi(M) ^ 0, namely those
occurring as factors of the annihilator Ann(T(M)) of T(M).
Copyrighted Material
186 Chapter 12. The Decomposition of Modules
C = {xi\i£Z}
To see that these groups are really Z and Z n in disguise, note that there
are bijective maps
:
Qoo ^ —t ^ o o j
given by
aoo(i) = x\
and
given by
an(i) = x\
12.11. Lattices 187
and
12.11 Lattices
Our results on additive groups have a geometric interpretation when we g
consider finitely generated additive groups that are subgroups of a real
vector space R™. Such an additive group L is called a lattice in R n . It
is clear that L must be torsion-free as a Z-module. Thus, by Theorem
12.3.1, L is isomorphic to Z7" for some integer r, and hence L has a basis
{ a i , . . . , ar}. We say that L is a full lattice in R" if r = n, in which case
{ a i , . . . , an} is also a basis of R n .
The reason for the use of the term lattice can be seen from the following
diagram in R 2 :
188 Chapter 12. The Decomposition of Modules
• o o o c) O • o 0 0 0 0
o o o o 1» o o o o 0 • 0
o 0 • o <) O 0 o • 0 0 0
• o 0 o () O • 0 o 0 0 0
o 0 o o (» o o 0 0 0 • 0
o o 6i^ o c) O o o • 0 0 0
• o 0 c) o o 0 0 0 0
»-* Q Q Q —0 0
G— ©— © w h-i^™ V
0 6ei
o o • o c) O o o ^•62 0 0 0
• o o o c) O • 0 0 0 0 0
o o o o <» o o o 0 0 • 0
o o • o () O o o • 0 0 0
Here, we take L to be the lattice given by the standard basis {ei, 62}, and
M is a lattice with basis {61, 62}, where 61 = I I and 62 = I
The points belonging to M are indicated by solid circles • and those in L
by open circles o, except where they are hidden under a point of M.
A basis { a i , . . . , a n } of a lattice L defines a fundamental parallelepiped
II(L) whose vertices are the origin, the vectors a%,... ,an, and all the sums
it! + ■ •• + Otfei with i\ < • •■ < ik and 1 < k < n. We then associate with
L the volume
vol(L) = I det(ai . . . a n ) |
of II(L), which is (by definition) the absolute value of the determinant of
the matrix whose columns are the (column) vectors a\,... ,an.
For n = 2, we prefer to speak of parallelograms and area. In our illus
tration, a fundamental parallelogram II(L) of L is indicated by dotted lines
and a fundamental parallelogram II(M) of M by solid lines. It is easy to
see that L has area vol(L) = 1, while vol(M) = 6.
Our main result in thi$^§^$gyj^p j^p^f^he volume of a lattice relates
12.11. Lattices 189
to the volume of a sublattice. We write \G\ for the order of a finite group
G.
12.11.1 Theorem
Suppose that L and M are both full lattices in the real space W1, and
that M is contained in L. Then the following statements hold.
(i) vol(L) does not depend on the choice of a basis for the lattice L.
(ii) The quotient group L/M is finite, and
vol(M) = vol(L)-|L/M|.
Proof
(i) Let { a i , . . . ,an} and {6i,... ,bn} be bases of L as a Z-module, write
det(P) = ± 1 .
Now let A = (ai ... an) and B = {b\ . . . bn) be the matrices whose columns
are the vectors in the respective bases. A careful check shows that
B = AP,
which explains why our scalars have suddenly appeared on the right, and
also why we had an unexpected transposition of suffices in our original
definition of the change of basis matrix.
Thus
det(B) = ±det(A),
which proves the assertion.
(ii) By Theorem 12.1.1, there are bases of L and M of the forms {bx,..., bn}
and {S\bi,..., 5rbr} respectively, where S\,... ,5r are the invariant factors
of L/M. But a basis of M is also a basis of R n , so we must have r = n. By
Theorem 12.3.1, we have
L/M^Z6l x.-xZ,n,
Co$ffi8t\tedWlatwfal
190 Chapter 12. The Decomposition of Modules
Clearly,
vol(M) = |det(5i&i . . . §
= <$i...(5n|det(&i . . . bn)\
= \L/M\vo\(L).
D
Thus, in the example above, we have L/M = 1,6. It is easy to see that
the basis {62, ^1} is a basis of L and that {62,6ei} is a basis of M.
Exercises
12.1 Let M = Z a x Z with a / 0 . Write mx = (T, 0) and m 2 = (0,1). Show
that M = Zmi © Zm 2 and that Z a S TLrax and Z £ Zm 2 .
Find all elements x = x\m\ + X2W2 € M so that Z = Zx.
For which of these x is there an element y G M so that M =
Zx © Zyl
Hint: consider intersections first.
12.2 Describe the invariant factor and elementary divisor forms of the Z-
modules with the following presentation matrices.
2 4 10 \
(a) 2 4 11
\ 3 7 0/
2 '\
(b) (I
I 32 12 ')
f \
(c) I 33 6 I.■ (See
(s Exercises 11.2 and 11.3.)
\2 a /
Exercises 191
/ 1 0 a \
12.3 Let A = 0 1 a I be a matrix over a field F, and let M be the
\ 0 0 b )
F[X]-module defined by the action of A on F 3 .
Discuss the possibilities for the invariant factor and elementary
divisor forms of M as a and b vary. (See Exercise 11.4.)
12.4 The challange again! Using Exercise 11.5, find the invariant fac
tor and elementary divisor forms of the Z[i]-module with generators
mi,m2,m3 and relations
Our aim in this chapter is to describe some normal forms for a square
matrix A over a field F. Before we can begin, we must describe what we
are seeking. A normal form for A is a matrix C whose entries conform to
some standard pattern and which is similar to A, that is, C = PAP~l for
an invertible matrix P over F.
The first normal form that we find is the rational canonical form. This
form can be computed from the invariant factors of the characteristic matrix
XI — A, and so it can be found by purely algebraic calculations, by which
we mean computations that involve only addition, multiplication and long
division of polynomials. The calculation of rational canonical forms enables
us to solve the similarity problem, that is, we can determine precisely when
two matrices are similar.
The second form is the Jordan normal form. This form is more elegant
when it can be obtained, but it can be found only when all the irreducible
factors of the characteristic polynomial det(XI — ^4) are linear polynomials
X — A. Thus, if we wish to ensure every matrix over F has a Jordan normal
form, we must impose the condition that every irreducible polynomial over
F is a linear polynomial, or, in other words, that F is algebraically closed.
Even when this requirement is satisfied, there is in general no purely alge
braic method for determining the Jordan normal form, since the roots of a
polynomial cannot be determined algebraically.
We also discuss the versions of the Jordan normal form that can be
found when the irreducible factors of the characteristic polynomial need
not be linear.
193
194 Chapter 13. Normal Forms for Matrices
A' = PAP'1
where P = PB\B is the change of basis matrix (Eq. (5.13) of section 5.11).
Conversely, if we are given that A' is similar to A through P , then we
choose B' so that P = PB',B- Thus A' is again the matrix of the action of
X with respect to the basis B'.
Of course, these two types of change are really two interpretations of
one phenomenon. If we are given a basis B = {b\,..., bn} of Fn and an
isomorphism 7r : Fn -)■ Fn, then B' = {7r(6i),... ,7r(6„)} will be another
basis of Fn. Vice versa, if B' = {b'j,..., b'n} is another basis, we can define
an isomorphism TX by ■K<^@pyr^m^l4terial
13.2. The minimum polynomial 195
These remarks suggest our strategy for finding a normal form for A: we
analyse the structure of M(A) as an F[X]-module and hence choose a basis
of Fn so that the action of X is represented by a matrix in some desirable
form. We start by reviewing the results that we found in previous chapters.
As we saw in section 9.7, the characteristic matrix
T= XI-A,
is a presentation matrix for M(A). The n-th Fitting ideal of XI - A is
generated by the characteristic polynomial det(XI — A) of A, which is a
monic polynomial of degree n (section 11.4). The invariant factors of XI-A
are polynomials Si(X),..., 6n(X) that satisfy the relations
61(X)... 5n(X) = det(XI - A), S^X) | • • • | Sr(X),
and, by Corollary 11.2.2,
51(X)...5h(X)=Fith(XI-A) for h=l,...,n.
Since the nonzero constants are units in F[X], we can take the invariant
factors to be monic polynomials, and then, by Theorem 11.3.1, they are
uniquely determined by the matrix XI—A and hence by the matrix A itself.
We also know from Corollary 12.8.2 that the invariant factors are uniquely
determined by the module M(A), that is, no alternative presentation of
M(A) can give different invariant factors.
Thus, by Theorem 12.3.1, M(A) has an invariant factor decomposition
M(A) £ F[X]/F[X]S!{X) x • • • x F[X]/F[X]5n(X) (13.1)
as an F[X]-module.
Note that the rank w of M{A) must be 0, that is, M(A) cannot have a
nontrivial free component F[X]W. The simplest way to see this is to note
that M{A) is finite dimensional as a vector space over F and F[X] is not.
A more sophisticated approach is to recall that Ann(M(A)) ^ 0 since it
contains the characteristic polynomial det(XI — A) of A (see Exercises 9.1
and 9.2).
In general, some of the invariant factors of XI — A will be the constant
polynomial 1. As these invariant factors give a zero component of M(A),
we say that they are trivial invariants factors. We sometimes omit such
terms from expressions such as that in Eq. (13.1) above.
13.2.1 T h e o r e m
(i) Ann{M(A)) = F[X]6n(X).
(ii) Sn(A) = 0.
(Hi) If h{X) is any other polynomial with h{A) 0, then 5n(X) divides
h(X).
0 0 • • 1 0 — fn-2
V0 0 • • 0 1 — fn-1 )
be the companion matrix of the polynomial / = fo + fiX-\ h/n-i^n_1 +
Xn.
Then, as in Proposition 12.4.1,
C6p$rl&)te<¥Matarial^ d
13.3. The rational canonical form 197
M = Mi 0 • • • © M r
with
Mi*F[X]/F[X)Si(X) for i = l,...,r.
Choose a basis Bi of each summand M, as a vector space over F, and let
Di be the matrix representing the action of X on M% for each i. Then the
union B = B\ U . . . U Br is a basis of M, and, as in section 7.5, the matrix
of the linear transformation X with respect to this basis is a block diagonal
matrix
/ £>i 0 . . . 0 \
0 D2 ... 0
D dia.g(Di,...,Dr).
\ 0 0 ... Dr /
Fix an index i. We exploit the fact that we have an isomorphism
Bi-.Mit* F[X]/F[X]Si{X)
Bi = {&i,i, ■ ■ • ,bi<n^}
where
»(i)-i
0(&i,0tw>^ft?«f(M$$at x
198 Chapter 13. Normal Forms for Matrices
and n(i) is the degree of St. Then the corresponding matrix A for X is the
companion matrix C(6i) of Si. Put
1
Sl = *W + 5 « X + ... + 5 « X^" + Xn»,
so that
0 0
(° 0
0
..
.. 0 0
-4° \
1
C(*i)
0 1 .. 0 0
-4°
0 0 . 1 0 °n(t)-2
0 .. 0 1 °n(t)-l /
We summarize these remarks, and a little more, as a theorem.
is a block diagonal matrix over F, in which the diagonal terms are the
companion matrices C(Si) of the nontrivial invariant factors Si of the char
acteristic matrix XI — A of A, and
<5il...|<5r.
Proof The points not covered by the preceding discussion are the existence
of the invertible matrix P and the claim about the invariant factors.
In the notation of section 5.11, A = (X)E,E is the matrix of the linear
transformation X of Fn with respect to the standard basis E of F n , while
C(A) = (X)B,B is the matrix of X with respect to B. Let P = PB,E be
the corresponding change of basis matrix, which is invertible, with inverse
P~l = PB,B- Then Eq. (5.13) gives C(A) = PAP'1 as required.
Since A is similar to C(A), we see that XI — A is similar to XI — C(A),
so both matrices have the^sa^JnyMianl.faGtprs by Theorem 11.3.1. □
13.3. The rational canonical form 199
13.3.2 Corollary
The rational canonical form of A is unique.
Proof
Suppose that a matrix A has two rational canonical forms, say
PAP-1 =C = diag{C(51),...,C{5r))
QAQ-1 = C = d i a g ( C ( ^ ) , . . . , C{5'r,))
with
5'x\...\5'r,.
Since the matrices C and C" are similar, their characteristic matrices
XI — C and XI — C" must have the same invariant factors, again using
Theorem 11.3.1. We know that the nontrivial invariant factors of XI — C
are the same as those of XI — A, namely 5\,..., Sr, the remainder all being
the identity 1. However, a direct computation of Fitting ideals shows that
the nontrivial invariant factors of XI — C must be 5[,..., 6'r,. It follows
that r = r' and, since all these polynomials are monic, that 5, = 5[ for
i = l,...,r. □
We can now determine when two nxn matrices A and A' over the field
F are similar.
13.3.3 Theorem
Two nxn matrices A and A' over a field F are similar if and only if
their characteristic matrices XI — A and XI — A' have the same invariant
factors.
Proof
If A and A' are similar, so are their rational canonical forms. By Corol
lary 13.3.2, these rational canonical forms must be the same, so that the
invariant factors of the matrices XI — A and XI — A' must also be the
same.
Conversely, if XI — A and XI — A' have the same invariant factors, A
and B are both similar to the same matrix in rational canonical form and
so are themselves similar.
□
200 Chapter 13. Normal Forms for Matrices
w1 = 1,
VJI = (X - X)w\,
u>3 = (X - A)w2 = (X-A)V, (13.2)
Wt (X-X)wt-! = (X-A)*"1^.
/ A 0 0 0 0 0 \
1 A 0 0 0 0
0 1 A 0 0 0
J(X,t) (13.4)
0 0 0 A 0 0
0 0 0 1 A 0
\ 0 0 0 0 1 A )
is a block diagonal matrix over F, in which the diagonal matrices are ele
mentary Jordan matrices.
The size t(i,j) of the entry J{Xi,t{i,j)) is given by the exponent of
X — Xi in the factorization
M = Mi,! © • • • © Mi,j © ■ • • © M s , r
in which
AfijeiF[X]/F[X\{X->n^,
the polynomials
13.4.2 Corollary
Let A be annxn matrix over a field F and suppose that A has a Jordan
normal form. Then the Jordan normal form of A is unique, apart from the
order in which the elementary blocks are written.
Proof It is clear that we can always permute the order of the diagonal
blocks of the Jordan normal form, since this amounts to a renumbering of
the roots of the characteristic polynomial of A. The argument to show that
no other change is possible parallels that given for the rational canonical
form in Corollary 13.3.2 above. If J = J(A) is a Jordan normal form of
A as constructed in the theorem, then the nontrivial invariant factors and
hence elementary divisors of XI — J must be the same as those of XI — A.
However, direct calculation shows that the invariant factors correspond
ing to an elementary Jordan matrix J(A, t) are 1 , . . . , 1, (X — A)*, which are
also the elementary divisors of XI — J(\,t). Thus, if
H-ll)
has characteristic polynomial X2 + 1, which is irreducible over the real
numbers.
This failing can be remedied in two ways. Given a polynomial f(X) over
a field F, it is always possible to embed F in a splitting field E in which
all the irreducible factors of f[X) are linear (Proposition 2.13.1). We can
therefore construct a field over which A does have a Jordan normal form
by adjoining roots of the characteristic polynomial of A to F.
Thus if we adjoin i = y/—l to R, obtaining the complex numbers C,
then the matrix A above has the Jordan normal form
o - "
The second, more drastic, method is to embed F in an algebraically
closed field E. By definition, E has the property that the only irreducible
polynomials over E are the linear polynomials X — A. This property can be
restated as "every nonconstant polynomial over E has a root in E". It is
a fact that any field can be embedded in an algebraically closed field - see
§6.1 of [Cohn 2].
The field C of complex numbers is the most familiar example of an
algebraically closed field, but some analytic tools are required to establish
this fact. A proof is given in §7.4 of [Cohn 2]. Moreover, the evaluation in C
of the roots of a polynomial over Q (for example) cannot be always carried
out algebraically, as is evidenced by the existence of quintic polynomials
that cannot be solved by radicals.
Thus the Jordan normal form cannot be computed purely by algebraic
computations in the polynomial ring F[X] save in special circumstances.
Notice that the rational canonical form of a matrix A is computed
through the invariant factors of the module M(A), but that the Jordan
normal form requires instead the elementary divisors of M{A).
dan normal form to distinguish it from the standard version of the Jordan
normal form.
As usual, let M(A) be the F[.X"]-module associated to a square ma
trix A. The primary decomposition of M{A) as F[X]-module expresses
M(A) as a direct sum of components which are (isomorphic to) cyclic mod
ules F[X]/F[X]p(X)k, where p(X) varies through the irreducible factors
of XI -A.
On general principles, A is similar to a block diagonal matrix J+{A)
whose diagonal terms correspond to the action of X on the various cyclic
components of M{A). We shall give a description of a typical diagonal term,
which we designate J+(p, k).
It will be convenient to use a double-suffix notation to describe the basis
that we construct. Write p(X) = p0 + p\X -\ 1- Xh, so that deg(p) = h,
and put
u>i,i =T,wi, 2 = X w i , i , . . . , w u = X h _ 1 w;i ? i.
Notice that if we reduce mod p, these elements map to the canonical F -
basis 1,lC, ...,Xh~l of F[X]/F[X]p(X) as constructed in section 2.12.
Now for i = 2 , . . . , k we define
F{x}p(xy-l/F[x}p(xy s F[X]/F[X\P(X),
the collection W = {wij} spans M as an F-space. Since W has kh mem
bers, it is therefore an F-basis of M.
The action of X on the basis elements is as follows. For j < h,
Xwij = wi]J+i
(1,1), ( 1 , 2 ) , . . . , (1, h); (2,1), (2, 2 ) , . . . (2, h);...; (k, 1), (k, 2 ) , . . . , (fc, h).
Thus the rows and the columns of J+(p,k) must also labelled with
double-indices (i,j) arranged in this order.
The column (i,j) of J+(p,k) gives the effect of X on Wij. For j < h,
we see that the (i,j)-th column is
(.. . , 0 ; 0 , 0 , . . . , 0 , 1 , 0 , . . . , 0 ; 0 , . . . ) T
where we exhibit the entries in the rows labelled (i, 1 ) , . - . , (i, h) (we have
transposed the column for convenience). The only nonzero entry is in the
(i,j + l),(i,j)-place.
For j = h and i < k, the (i, h)-th column is
the entry — pi occurs in row (i, h) while the entry 1 is in row (i + 1, h).
Finally, for j — h and i = k, column (k, h) is
(•■•,0;-p0,-Pi,- -Ph-i)
(o 0
0
• ■ 0
■ • 0
-Po \
1 -Pl
C = C[p)
0 0 • • 0 -ph-2
0 • • 1 -Ph-1 I
/ 0 0 0 1\
0 0 0 0
Y =
0 0 ••• 0 0
Cor)yifyh&d Material0 J
206 Chapter 13. Normal Forms for Matrices
Then
/ C O O 0 0 0 \
Y C 0 0 0 0
0 Y C 0 0 0
J+(p,k) (13.5)
0 0 0 c o o
0 0 0 Y C 0
\ o o o 0 Y C )
cI 0 0 • 0
■■ 0 0
c 0 •• • 0 0 0
0 I c ■■ ■ 0 0 0
J°(p,k) (13.6)
0 0 0 •• • c 0 0
0 0 0 •• ■ I c 0
\ 0 0 0 •• ■ 0
cC ) I
where C is the companion matrix of p and I is an h x h identity matrix,
h = deg(p). Such a block has the same form as the matrix J+(p, k) of Eq.
(13.5) above except that the matrix Y is replaced by an identity matrix I
throughout.
The matrix J s (^4) is called the separable Jordan form for A.
The claim will be established by showing that Js(p,k) is similar to
+
J (p, k) for any p and k. Since the polynomialp has distinct roots Ax,..., A/,
in the extension field E, we know that the Jordan normal form of the matrix
Cis
/Ax 0 ... 0 \
0 A2 ■•• 0
-l
A SCS
\ 0 0 ■■■ A,, /
where S is an h x h invertible matrix having entries in E.
Let T = diag(5 S ■ ■ ■ S) be a block diagonal matrix with k copies of S
on the diagonal. Then
A 0 0 • • 0 0 0
(
J A 0 • ■ 0 0 0
0 I A • ■ 0 0 0
1
J' = TJ°(p,k)T- = (13.7)
0 0 0 ■ ■ A 0 0
0 0 0 ■ • / A 0
^o 0 0 ■ • 0 I A /
and
Next, we permute the rows and columns of J' by moving row (i — l)h+j
to row (J — l)k + i and likewise moving column (i — l)h + j to (j — l)fc + i,
obtaining a new matrix J".
Let P be the hk x hk permutation matrix corresponding to the permuta
tion of the rows, that is, P is the result of performing the row permutations
on an hk x hk identity matrix. Then the matrix P ~ x corresponds to the ef
fect of performing the column permutations on an hk x hk identity matrix,
and
J" = PJ'p-\
We now observe that the nonzero entries of J " are
and
A 0 0 • • 0 0 0
( J \
1 \j 0 • ■ 0 0 0
0 1 A, • • 0 0 0
Ai =
0 0 0 • A, 0 0
0 0 0 • 1 Xj 0
Vo 0 0 ■ ■• 0 1 ^ 1
which is evidently the elementary Jordan matrix associated to the poly
nomial (X — \j)h- Hence J " is in Jordan normal form, associated to the
polynomial
(X-\1)k...(X-\h)k=p(X)k.
But then J" must be the Jordan normal form of the matrix J+(p,k),
+
since the characteristic matrices of both J" and J (p,k) have the same
invariant factors. It follows that J " and J+ (p, k) are similar, and, trac
ing through the various similarities that we have used, that J" (p, k) and
J+(p, k) are similar as ™%mm®<fitiaterial
13.8. Nilpotent matrices 209
Finally, we have to prove that Js(p,k) and J+(p,k) are in fact similar
as matrices over F, that is, we can find a matrix Q with entries in F so
that QJs{p,k)Q~l = J+(p,k).
As matrices over E, the characteristic matrices of both J"{p,k) and
J+(p,k) have the same set of invariant factors. However, these invariant
factors already belong to F, since both matrices in fact have entries in F,
which shows that Js(p,k) and J+(p,k) are similar over F, by Theorem
13.3.3.
Exercises
13.1 Let
-3-2 4 \
'■[ 4
0 - 1
1-4
1 /
be a complex matrix. Find the rational canonical form and Jordan
normal form of A, and show that they are the same as for
-1 0 0
B 0 0 -1
0 1 -2 )
Notice that B is not in rational canonical form, although it is a
diagonal block matrix made up of companion matrices. Explain.
13.2 Find the rational canonical form and Jordan normal form for each of
the following matrices:
2 0 0 0 \
1 0 1
1°
A=| 0 0 M
1 ; B = ! o o 1
0
\
C
3 2 0 -2
0 0 2 0
Vo 0 0/ l l 0 o/ V0 0 2 2
and
/ 1 1 1 • • 1 1
0 1 1 • • 1 1
D 0 0 1 • • 1 1
V\00 00 00 ■■•■■ 00 11 / /
where D is an n x n matrix.
13.3 Repeat Exercises 13.1 and 13.2, but regarding the matrices as having
entries in the following fields in turn: R, Z2, Z3, Z5. (So this is 6 x
4 problems in one. You should always find a rational canonical form,
but there may be no Jordan normal form.)
Find the nonspliC<^9/dgfi^6rfiMa/6gJS/there is no Jordan form.
212 Chapter 13. Normal Forms for Matrices
13.4 Let A be a square matrix over a field F, and let AT be its transpose.
Show that the invariant factors of the characteristic matrices XI - A
and XI — AT axe the same. Deduce
(a) A is similar to AT;
(b) the F[X]-modules M(A) and M(AT) are isomorphic.
13.5 Let F be any field. Find the possible Jordan normal forms of an n x n
matrix A which satisfies the equation A2 = A.
s 13.6 Let p(X) be a separable irreducible polynomial over a field F , with
p(X) = (X-X1)...(X-Xh)
for i = 1 , . . . ,k.
Let J' be the matrix in Eq. (13.7), and let Mi be the (hk - 1) x
(hk - l)-minor of XI - J' formed by eliminating row 1 and column
h+1 of J'.
Show that Mi = p1.
Find (hk - 1) X (hk - l)-minors Mi of XI - J' with M{ = pi for
each i.
Deduce that Fithk-i(J') = 1, and devise an argument to show
that J' is similar to J+ (p, k) without using the ancillary matrix J " of
section 13.7.
s 13.7 This exercise shows that the hypothesis of separability is essential for
the equivalence of the separable Jordan normal form with the nonsplit
version of the Jordan normal form.
Let Q be any field and let Q[t] be the ring of polynomials over Q
in an indeterminant t. Further, let F = Q(t) be the field of rational
functions over Q, that is, the field of fractions of Q[t\. Verify that the
polynomial p(X) = X2 — t is irreducible over F.
Let E = F(y/t) be a splitting field for p(X). If the characteristic
of Q is not 2, that is, 2 ^ 0 in Q, then p(X) = (X + Vt)(X - yfi)
is separable, but if Q has characteristic 2, say Q = Z2, then p(X) =
(X — v^) 2 and p(X) is not separable.
/ 0 t 0 0 \
1 0 0 0
Let A which is the matrix Js(p,2) of Eq.
l o o t
\ 0 1 1 0J
(13.6).
Exercises 213
Show that
F i t i p f l - A) = Fit2(XI - A) = 1,
F i t 3 ( X I - A) = p(X)
Projective Modules
In this final chapter, our aim is to provide some contrast to the results that
we have obtained for the structure of modules over Euclidean rings. We do
this by taking a brief look at those modules which occur as a direct sum-
mand of a free module - these are the projective modules. If the coefficient
ring is Euclidean, then a projective module must be free, since any sub-
module of a free module is free. For rings in general, a projective module
need not be free, nor need a submodule of a free module be projective.
We also discuss the types of ring that are defined by imposing "pro-
jectivity" conditions on modules, and we show that one such class, the
Artinian semisimple rings, occurs naturally in the representation theory of
groups.
The material in this chapter is all supplementary to the original lee- g
ture course on which these notes are based. Besides the aim of placing
the Euclidean results in a wider context, the results and references given
here provide an introduction to some topics that might be included in an
"enhanced" MMath or MSci version of the course.
P x Q S R1.
215
216 Chapter 14. Projective Modules
.EaiampZes.
(i) It is immediate from the definition that the ring R is itself a projective
left (and right) R-module, as is any free module R1.
(ii) The zero module is projective.
(hi) Let R be the ring of all n x n matrices over a field F , and for j = 1 , . . . ,n,
let Ij be the set of matrices whose entries are all 0 except in column j .
Then (Exercise 7.8) each Ij is a left H-module and
R = h ©-•■«!„.
Thus each summand Ij is projective. Note that the summands are all
isomorphic to one another as left .R-modules.
No Ij is a free .R-module, since dim(Jj) = n as a vector space over F,
while the dimension of any free R-module is a multiple of n 2 .
(iv) Let D be the ring of 2 x 2 diagonal matrices over a field F, and let e = en
and / = / n (see Exercises 1.7 and 7.4. Then D = De®Df', which shows
that De and Df are (nonisomorphic) projective D-modules. It is easy
to see that neither is a free D-module.
(v) The cyclic Z-module Z a , a > 1, is not projective. To establish this
fact, we argue by contradiction. Suppose that there is an isomorphism
6 : Z a x Q = Z1 for some module Q and index set I. The image of (1,0)
is a torsion element in Z 7 , since it is annihilated by a, and it is nonzero
since 6 is an injection. But Z 7 contains no nonzero torsion elements.
(vi) Let A be an n x n matrix over a field F, and let M be the F[X]-module
defined by X acting as A on Fn. Arguing as above, we see that M is not
projective as an .F[X]-module.
Before we give more examples, it will be useful to reformulate the defi
nition in terms of the splitting of homomorphisms.
14.2.1 Lemma
Let M and P be left R-modules and suppose that n : M —► P is split by
a : P -> M. Then
14.2.2 Theorem
Let P be a left R-module over a ring R. Then the following statements
are equivalent.
218 Chapter 14. Projective Modules
Proof (ii) => (i). Let {pt \ i E 1} be any set of generators of P, where the
index set I may not be finite. An element x = (XJ) of the free module R1 is
a sequence of members Xi of R, indexed by / , and with only a finite number
of nonzero terms. Thus we can define a surjective .R-module homomorphism
9 : R1 —> P
by
6{x) = ^XiPi.
iel
Since 9 is split, P is projective.
(i) => (ii). By definition, we have
R1 ^PxQ
for some index set / and module Q. Let 9 : R1 —> P be the corresponding
surjective homomorphism and let ui : P —> R1 split 9. Let {ej} be the
standard basis of R1 (Exercise 7.10) and put pi = 9(ei) for each i, so that
{pi} is a set of generators of P.
Since n : M —> P is surjective, we can choose a set of elements {rrii} of
M so that
7r(m;) = pi for all i £ I.
Now define A : R1 —> M by the requirement that
A(ej) = rrii for all i e i ,
and write
a = \u>.
Then
7Tcr = TTXU) = OUJ = idp,
14.2.3 Corollary
Let P be a left R-module over a ring R. Then P is a finitely generated
projective module if and only if there is a split surjection
R* —>P
14.2.4 Corollary
Suppose that R is a Euclidean domain. Then every finitely generated
projective R-module P is free.
2 • 3 = (1 + V^5) • (1 - v 7 ^ ) ;
u>(z) --
z I
1
An easy calculation confirms that the image of a? is in R?2 and that 9ui = idj.
E x a m p l e : a non-projective ideal.
Let R = F[X, Y] be the polynomial ring in two variables over a field F,
and let I = RX + RY be the ideal generated by the variables. It is easy to
show by direct calculation that / is not principal, and we now show that /
is not projective. This result provides a contrast to the fact that any ideal
of a Euclidean domain is principal, and hence free.
There is an evident presentation
9:R2^I
with
9(f,g)Dep^figftt9^IVfate^klf,9 € R.
220 Chapter 14. Projective Modules
Similarly,
w(F) = (c,d)
with
n n—1
c = ^ C j ( X ) y ^ and d = I - d{X)X ~^2cj+1{X)XY:>.
3=1 3=1
Yb = Xd
CopyrightedMaterial
14.3. Semisimple rings 221
where each I\ is a minimal left ideal of R, and A is some index set that
need not be finite.
The first result is as follows.
14.3.1 Theorem
A ring R is left semisimple if and only if every left R-module is projec-
tive.
Proof In one direction, the argument is not too difficult. A matrix ring
Mn(F) over a field F is Artinian since it is a finite dimensional vector space
over F, and any ideal is a subspace. Exercises 3.10 and 7.8 combine to show
that Mn(F) is semisimple. It is not hard to see that essentially the same
arguments work when F is replaced by a noncommutative division ring D
and that a direct product of rings is Artinian semisimple if its components
are.
The proof in the reverse direction is much harder. Full details can be
found in [Cohn 2], §4.6 or [B & K: IRM], §4.2. □
The structure of modules over an Artinian semisimple ring is transpar
ent. For each matrix ring Mi(Di) above, let Si be its "first column", that
is, the set of matrices whose entries must be zero outside the first column.
Then Si is a left Mi(Di)-module and we can make Si into left R-module
by stipulating that the other components of R act trivially on Si. Proofs
of the following result c a ^ b e ^ u ^ ^ j h ^ ^ f ^ r e n c e s given above.
222 Chapter 14. Projective Modules
14.3.3 T h e o r e m
Let M be a finitely generated left module over an Artinian semisimple
ring R. Then there are non-negative integers ai,...»ftfe and an R-module
isomorphism
M = oiSi x • • • x akSk,
where a^Sj denotes the external direct sum of a^ copies of Si. (If some
Oi = 0, we take this to be the zero module.)
If N is a finitely generated left R-module with
TV^feiS1! x ••• xbkSk,
then
M = N <=s>
^=> ai = bi for i = \,...,k.
U
xy
x
y== J2( 22 xhxxhXi)gj.
^2( Yl i)9i-
3= 1
24.4. Representations of groups 223
C = ( c ) = {l,c,...,Cfc-1}
(0(gh))E,E = (6(g))E,E(9(h))E,E
Exercises
14.1 Let R = F[X, Y] be the polynomial ring in two variables over a field
F, and regard F as an .R-module with both X and Y acting as 0.
Show that the obvious presentation R—>F has kernel / = RX + RY,
226 Chapter 14. Projective Modules
R2 ->L
(0 ^Xf + Yg,
FC = F[e] S F[X]/F[X]X p .
229
230 Hints and Solutions
4200 = 2 3 • 3 • 5 2 ■ 7 = -i • (1 - if ■ 3 ■ (2 + i) 2 ■ (2 - i)2 • 7.
which gives PD,c • Pc,B = PD,B by the result above. (Alternative proof:
calculate the product explicitly!)
5.6 Prom the above, PD,B = -PD.B-PE.BJ where
2a — 2 — i = 1, — l , i o r — i.
This gives
a = (3 + i)/2, (1 + i)/2,1 + i or 1.
But the first two of these are not allowed, since they are not Gaussian
integers, so the permitted values are a = 1 + i, 1.
5.8 Similar to the preceding exercise: we determine the polynomials
f(X) for which the matrix P = ( _ yiA j is invertible over R[X]. The
determinant is f(X) — 2g(X), and the units in R[X] are the nonzero real
numbers, so the permitted values of f{X) are r + 2g(X), r / 0 e R; g(X)
is arbitrary.
6.1 The initial parts are routine checking.
An element of Hom(Z p , Z q ) is determined by an integer r such that pr € Zq.
This means that q divides pr, and so q divides r since p, q are distinct prime
numbers (Proposition 2.8.2). Hence f = 0 in Z q , so the corresponding
homomorphism must be 0.
Since Z p (p) = Z p (p 2 ) = Z p , we have
Hints and Solutions 233
( X 0 0 0 /o \
X 0 0 /i
0 o h
XI-A
0 0 -1 ^ /n-2
Vo 0 0 -1 /n_!+X /
0 0 0 f(X) \
f °
-1 0 0 0 /i + - + /„-iI"- 2 +X'- 1
0 -1 0 0 f2 + ■ ■ ■ + fn-lXn~3 + Xn~2
B =
0 0 -: 0 / n _2 + / „ - l X + X 2
0 0 -1 fn-l+X
Io
Expanding from the first row, det(B) = ( - l ) " " 1 / • ( - l ) n _ 1 = / .
6.6 In these pictures, "x" denotes the submodule xM of M.
(c): R/Rlpq •
Ip
0 •
234 Hints and Solutions
(d): R/Rp2q*
0 •
(e): R/Rp2q2 „ 2 ,
pq
p2q pq2 •
0 •
Hints and Solutions 235
RPi...pk-i/Ra^R/Rpk,
N^F[X}/F[X](XZ-1)
for any field F, which can be seen by a slight modification of the argument
in section 6.6.
Since X3 — l = (X — l)(X2+X+l) with the second term irreducible, we have
N = P®Q with P = (X2 + X + 1)TV one-dimensional and Q = {X - 1)TV
two-dimensional.
Over Z 3 , X 3 - 1 = (X - l ) 3 and so TV has no direct sum decomposition -
its submodules form a chain.
7.3 Define
by
a(mi,...,mfe) = ( m l r . . , m t ) ,
and check that a is surjective with kernel L\ x ■ • • x Lk, so that the claim
follows from the Induced Sag^fi^epfcMsfefl^/
236 Hints and Solutions
Pi = 7 i i « i + - - - + 7 t j e t
by the cofactor Fjk, k a fixed index, and then sum to obtain the relation
- « - ( S -.)■
Then M = Z 2 x Z, generators mj = - m i - 3m 2 - m 3 and m 2 = ma-
ll.1 Fitj = 1, Fits = 3, Fit 3 = 18, confirming that St = 1, 52 =
3, 53 = 6.
11.2
(a) Fit 2 (r) = ± d e t ( r ) = 1; <5i<52 = 1, hence 5t = 52 = 1.
(b) Fiti = <5i = 1. Fit 2 has among its generators 2 x 1 1 - 2 x 1 0 = 2 and
4 x 0 - 7 x 1 1 so Fit 2 = 52 = 1 also. Expanding, say, from row 3, we get
det = - 2 so 53 = 2.
(c) Fiti = 1, Fit 2 = 3.
(d) Fiti = 1 and Fit2 is generated by 12,3o - 4 and 3a - 12, so that Fit 2
is given by the GCD of 3a, 4.
For a = 0, this is 4, for a = 1, this is 1, and for a = 2, this is 2.
11.3 Since the GCD of 3a, 4 must divide 4, the only possible values of
Fit 2 are 1, 2,4, so the values a = 0,1, 2 cover all the possibilities.
11.4 Fiti = 1 always. If a ^ 0, a(X — 1) is a 2 x 2 minor, and X — 1
is a factor of any 2 x 2 minor. So Fit 2 = X — 1, therefore 52 = X — 1, $3 =
(X - l)(X - b).
If a = 0, the distinct l x l minors are X — 1 and X — b, and the distinct
2 x 2 minors are {X - l ) 2 and (X - 1)(X - b).
If 6 ^= 1, we find that Fitj = 1, Fit 2 = X - 1 = 82 and 63 = {X - 1)(X - b)
again.
lib = l,61 = 52 = 53 = X-l.
11.5 The presentation matrix is
/ 0 6 6 0 \
T = l + i 4 5-i 2i \
\ 0 2(1 + *) 2£ 0 /
/1+i 0 0 0\
A= 0 2 0 0 ) .
\ 0 0 6 0/
This can be found by using the elementary row operations "row 1 -H- row
2", "row 2 «-> row 3" cleverly, or by finding the Fitting ideals of T.
Hints and Solutions 239
But for any given values of Xi and y2 we can find a value of 2/1 with ±2/2^1 +
2/i invertible in Z/Za. Thus x = (xi, ±1) is the general form for x.
12.2
(a) M = Z2 is 2-primary.
(b) M = Z 3 is 3-primary.
(c) For a = 0, M = Z 4 is 2-primary; for a = 1, M = 0; and for a = 2,
M is 2-primary again.
12.3
• If a / 0, the invariant factor decomposition is
12.4 By the calculations for Exercise 11.5, the invariant factor form of
M is
Z [ t ] / ( l + t ) Z [ i ] x Z[i]/2Z[t] x Z[i]/6Z[i].
We have irreducible factorizations 2 = —t(l + i)2 and 6 = - 1 ( 1 + i)2 ■ 3, so
the 1 + i-primary component of M is
Z[t]/(1 + t)Z[i] x Z[t]/2Z[t] x Z[i]/2Z[t]
and the 3-primary component is
Z[t]/3Z[i]
all other primary components being 0. The elementary divisors are
1 + *, 2,2,3.
In the following solutions, "InFs" means invariant factors, "RCF" rational
canonical form, and "JNF" Jordan normal form.
13.1 By direct calculation, the InFs of
/ X +3 2 -4 \
XI-A = 4 X -1 4
-
V o 1 X-l I
are 1,1 and {X - l)(X + I)2 = X3 + X2 - X - I.
/ 0 0 1 \ / 1 0 0 \
RCF: 1 0 1 , JNF: 0 - 1 0 .
\ 0 1 -1 ) \ 0 1 -1 /
B turns out to have the same InFs, so the same RCF, JNF. But B itself
is not in RCF: the l x l block is the companion matrix of the polynomial
X — 1 and the 2 x 2 block is the companion matrix of X2 + 2X + 1, which
is not a multiple of X — 1.
13.2
0 0 0 \
A: InFs 1,1, X 3 , RCF & JNF: : 0 0 I
\
\o i o)
B: InFs 1,1,Xs - 1, RCF:
( 0 0 1\
1 0 0 JNF:
1 0
w
°\ where
b 0 0
u) = exp(27ri/3).
Vo i o v° 0 2
- y
C: I n F s l , l , X - 2 , ( X - 2 ) 3
/ 2 0 0 0 \ / 2 1 0 0 \
0 0 0 1 0 1 0 1
RCF: JNF:
0 1 0 - 4 0 2 1 0
\ 0 0 1 4 ) Copyrigntefl IVfategiali
Hints and Solutions 241
(° 0 0. .0 /o
h
\ / 1 0 0...0 0 \
1 0 0 . .0 1 1 0...0 0
0 1 0 . .0 h 0 1 1 ...0 0
and
0 0 0 . .0 fn-2 0 0 0 ... 1 0
0 0 . .. 1 fn-1 ) \ 0 0 0 ... 1 1 /
/ A2 0 0 / A 0 0
2A A2 0 " ^ 1 A 0
2
1 2A A • 0 1 A
giving
A2 = A and 2A = 1.
These equations are incompatible unless the matrix J is 1 x 1, in which
case A = 0,1 work.
Ir 0
So, collecting all the l's together, the possible JNF's are for
0 0
identity matrices of size r = 0 , l , . . . , n .
No hints for Chapter 14-'
Bibliography
243
244 Bibliography
245
246 Index
of polynomials, 5, 29 sum, 42
as ED, 18 torsion, 124
complex, 26 zero, 41
over a finite field, 28 sum
rational, 27 of homomorphisms, 53
real, 27 of ideals, 9
of scalars, 36 of submodules, 42
of triangular matrices, 15 summand, 108, 112
quotient, 10 surjective, 59
residue, 10, 28, 29 Sylow subgroup, 187
trivial, 3 symmetric, 11
zero, 3
row & column operations Third Isomorphism Theorem, 103
general, 150 torsion, 124
row operations module, 124
elementary, 146 submodule, 124
torsion-free, 124
scalar matrix, 39 transitive, 11
scalar multiplication, 35 transpose matrix is similar, 212
Second Isomorphism Theorem, 103 trivial invariant factor, 195
semisimple, 220 trivial module, 40
separable Jordan normal form, 207 two-sided ideal, 7
separable polynomial, 206
set of generators, 44 underlying space, 55
similar matrices, 194, 199 Unique Factorization Theorem, 24
similarity problem, 193 unit, 4
simple module, 41 unit group, 5
Smith normal form, 152
vector space, 36
spanning set, 44
basis, 44
split homomorphism, 216
spanning set, 44
split polynomial, 31
volume, 188
splitting field, 31
standard basis, 60 Wedderburn-Artin Theorem, 221
standard factorization, 25
standard irreducible, 25 zero ideal, 7
standard matrix units, 120 zero module, 36
subgroup, 41 zero submodule, 41
submodule, 40
generated by set, 44
intersection, 42
of cyclic module, 96
proper, 41