MATH325-425-Representation-Theory-of-Finite-Groups-Notes
MATH325-425-Representation-Theory-of-Finite-Groups-Notes
[email protected]
http://www.maths.lancs.ac.uk/~grabowsj
1
Contents
0 Categories 4
1 Group actions 9
2 Group representations 11
3 Linear representations 15
3.1 Equivalence of representations . . . . . . . . . . . . . . . . . . 18
3.2 Subrepresentations . . . . . . . . . . . . . . . . . . . . . . . . 18
3.3 Quotient representations . . . . . . . . . . . . . . . . . . . . . 19
3.4 Direct sums . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
4 Algebras 20
4.1 Interlude—tensor products of vector spaces . . . . . . . . . . 21
4.2 Algebras, continued . . . . . . . . . . . . . . . . . . . . . . . 22
6 Modules 33
6.1 Modules for group algebras . . . . . . . . . . . . . . . . . . . 43
7 Character theory 47
A Burnside’s pα q β theorem 60
2
Introduction
This is a first course on representation theory, with a particular aim of
describing the representation theory of finite groups. We will introduce
the notion of group representations, to set the context for the shift from
representations—which have certain disadvantages—to modules. We give
the basics of the theory of modules for a general algebra over a field, and
then specialise to the group algebra of a group, this being an algebra that
carries all the information of the group. We will finish with the theory of
characters of a finite group, which is a more concrete way of studying specific
representations of groups.
We will introduce and use the language of categories, but not very much
of the theory of these. This language is that of modern representation theory
and allows us to make more precise statements.
The learning outcomes for this module are as follows. On completion of
the module you should
(c) know and be able to apply the main results on modules for group
algebras, such as Maschke’s theorem and Schur’s lemma, in particular
in the case of finite Abelian groups.
(d) know and be able to apply the main results on module homomorph-
isms, in particular in the case of the decomposition of the group algebra
of a finite group.
(e) know and be able to apply the main results on characters, such as
the orthogonality relations and construction of the character table of
a finite group.
3
0 Categories
In this preliminary section, we will introduce some new language that will
help us talk about representation theory. Categories are (as I hope you
will see) a very natural way to express relationships among collections of
algebraic objects, in a way that in particular respects functions. We will not
need very much actual category theory but using the terminology will allow
us to make better, more precise statements.
such that
(a) for each pair of morphisms f ∈ C(x, y) and g ∈ C(y, z), there exists a
composite morphism g ◦ f ∈ C(x, z) such that composition is associat-
ive:
h ◦ (g ◦ f ) = (h ◦ g) ◦ f
for all f ∈ C(w, x), g ∈ C(x, y) and h ∈ C(y, z); and
(b) for each x ∈ Obj(C), there exists an identity morphism idx ∈ C(x, x)
such that the left and right unit laws hold :
idy ◦ f = f = f ◦ idx
(a) There is a category with one object ∗ and one morphism id∗ .
e
(b) Let v → w be the directed graph Q with two vertices v, w and one
edge e. Then there is a category P(Q) (the path category of the graph)
with two objects, v and w, and P(v, v) = {idv }, P(v, w) = {e} and
P(w, w) = {idw }. If Q were any arbitrary directed graph, we would
take Obj(P) to be the set of vertices of Q and P(v, w) to be the set
of paths in Q from v to w. (Note that idv corresponds to the path of
length 0 from v to itself.)
1
Here and elsewhere, we will say “collection” due to what mathematicians call “size
issues”. These occur because the collection of all things we might be interested in might
not form a set, depending on our logical foundations. You might recall that issues such as
Russell’s paradox can occur if care is not taken. We will not dwell on this but will signal
that care is needed, by using “collection” instead of “set”.
4
(c) Let G be a group. Let G be the category with one object ∗ and
G(∗, ∗) = G. Note that eG = id∗ .
But, I hear you say, these examples are all a bit. . . unnatural. Yes: this
makes the point that the definition of a category includes a huge range of
structures.
As for some more familiar examples:
(d) Let Set be the category with objects Obj(Set) the collection of all
sets and morphisms Set(X, Y ) = Fun(X, Y ), that is, the collection of
morphisms between two sets X and Y is the collection of all functions
from X to Y . Then you have known for a long time that composition
of functions exists, is associative and that we have identity functions.
(e) Let Grp be the category with objects Obj(Grp) the collection of all
groups and morphisms Grp(G, H) = Hom(G, H) the collection of all
group homomorphisms from G to H. Composition and its associ-
ativity are inherited from Set and the identity function is a group
homomorphism.
(f) Let Rng be the category with objects Obj(Rng) the collection of all
(not necessarily unital) rings and morphisms Rng(R, S) = Hom(R, S)
the collection of all (not necessarily unital) ring homomorphisms from
R to S. Composition and its associativity are inherited from Set and
the identity function is a ring homomorphism.
(g) Let Ring be the category with objects Obj(Ring) the collection of all
unital rings and morphisms Ring(R, S) = Homu (R, S) the collection
of all unital ring homomorphisms from R to S. Composition and its
associativity are inherited from Set and the identity function is a ring
homomorphism.
(h) Let K be a field and let VectK be the category with objects Obj(VectK )
the collection of all K-vector spaces and morphisms VectK (V, W ) =
HomK (V, W ) the collection of all K-linear maps from V to W . Com-
position and its associativity are inherited from Set and the identity
function is a K-linear map.
(i) Let Top be the category with objects Obj(Top) the collection of all
topological spaces and morphisms Top(X, Y ) = C 0 (X, Y ) the collec-
tion of all continuous functions from X to Y . Some work is needed
to prove that this is a category! (Don’t worry if you don’t know this
example: it won’t appear again in this course, and this is why it was
included, i.e. to illustrate that there are examples that are natural but
not algebraic.)
5
Most of our categories will be like these, with morphisms being some
structure-preserving map. Going forward, our usual convention for morph-
ism sets will be to write HomC (X, Y ) for morphisms from X to Y in the
category C, to remind us to think of homomorphisms. When X = Y ,
we have a special name for the resulting morphisms; namely, we write
def
EndC (X) = HomC (X, X).
Just as sets, groups, rings, vector spaces etc. have structure-preserving
maps between them, so do categories. Indeed, it was claimed by Mac Lane,
one of the founders of category theory, that being able to define and study
these was the whole point. Indeed, just as bijections or group or ring iso-
morphisms or invertible linear maps encode for us when two sets, groups,
rings or vector spaces are “the same”, so certain maps of categories will do
this too.
f F Ff
x y ⇒ Fx Fy
g Fg
g◦f F (g◦f )
z Fz
2
In fact, it is one of two such, though arguably the other looks less natural at first
sight. Namely, what we have defined is called a covariant functor; its counterpart, a
contravariant functor, satisfies F (g ◦ f ) = F f ◦ F g.
6
Similarly, preserving identity morphisms means preserving loops.
Then the collection of all categories together with functors between them
forms a category. Unsurprisingly, we are running headlong into logical and
set-theoretical issues here again, but we can avoid some of this, as follows.
Exercise 0.6. Which of the above examples of categories are small? locally
small?
Definition 0.7. The category Cat with objects all small categories and
morphisms Cat(C, D) all functors from C to D is a category.
• F : Ring → Set
• F : Rng → Set
• F : Grp → Set
• F : VectK → Set
• F : Ring → Rng
• F : Ring → Grp
• F : Rng → Grp
• F : Top → Set
Most functors actually “do” something and we will see more non-trivial
examples later.
The first four of these and the last express that the categories in the
domain of F (i.e. our more familiar examples of categories) are what is called
concrete. The existence of these forgetful functors to Set is the precise way
to say that these categories consist of sets and functions with extra structure.
Especially in these examples, but also when we are generally feeling lazy,
we will often write x ∈ C to mean “x is an object of C” (i.e. x ∈ Obj(C)).
7
So for example we might say “let V ∈ VectK ” as a shorthand for “let V be
a K-vector space”.
We will conclude our whirlwind tour of “elementary category theory”,
if there is such a thing, by addressing the issue mentioned above of when
two categories should be considered to be “the same”. Perhaps surpris-
ingly3 , there are several levels of sameness, and the most direct analogue of
isomorphism is not the right one.
So the first idea we might have is to say that two categories C, D are the
same if there is an isomorphism F : C → D in Cat between them. That is, if
there are two functors F : C → D and G : D → C such that G ◦ F = idC and
F ◦ G = idD , where idC is the identity functor on C (sending every object to
itself and every morphism to itself).
However this is too strict a notion and isomorphisms of categories are
very rare. Instead, we should relax the conditions a little and say that G ◦ F
and F ◦ G should be almost the identity functors. To do this definition
properly requires the notion of a natural transformation of functors. For
expediency, we will give an equivalent definition that has the advantages of
both being easier to check and also more like the “injective and surjective”
condition we are used to in Set.
8
There is a special sort of equivalence, when the two functors have a
particular relationship, as follows.
1 Group actions
When you were introduced to groups, you were probably told that groups
often arise as symmetries. The subtle but important shift from “what group
encodes the symmetries of this object?” to “which objects does this group
give symmetries of?” moves us from the structure theory of groups (which
is what you have studied before) to their representation theory.
The word “symmetry” needs a little unpacking before we can make much
progress—we need some formal definitions in order to state and prove things.
Informally, a symmetry of an object should move the points of that object
around, there should be a symmetry that leaves the object unchanged and
if we have two symmetries, we should be able to apply one and then the
other to obtain another symmetry. Very often, we want our symmetry to
be reversible; asking for this is what puts us in the domain of groups.
The most common way to encode this formally is in the notion of a group
acting on a set, or a group action. We will also do this first, very shortly, but
we will then show how this can be transformed into different formulations
that are better suited for posing the questions we will want to ask.
4
This notation is shorthand for F : C → D, G : D → C and emphasises that we are
interested in F and G as a pair. By itself, it makes no claims about G ◦ F or F ◦ G,
however.
9
Definition 1.1. Let G be a group and X a set. A left action of G on X is
a function α : G × X → X such that
10
The orbitSof an element x ∈ X is a subset of X and the orbits partition
X, i.e. X = x∈X O(x) and O(x) ∩ O(y) = ∅ if x 6= y. The stabilizer of x
is a subgroup of G.
Exercise 1.7. Prove the claims in the preceding paragraph.
For a group G and subgroup H, we will write (G : H) = {gH | g ∈ G}
for the set of left cosets of H in G.
ϕ : O(x) → (G : Stab(x))
h . ϕ(y) = ϕ(h . y)
Proof. We have
g . x = g0 . x
⇐⇒ (g 0 )−1 . g . x = (g 0 )−1 . g 0 . x
⇐⇒ ((g 0 )−1 g) . x = x
⇐⇒ (g 0 )−1 g ∈ Stab(x)
⇐⇒ g Stab(x) = g 0 Stab(x)
Extension 1.9. How would you define a morphism between two G-actions?
That is, given α : G × X → X and β : G × Y → Y , what would be an
appropriate notion of a morphism f : α → β?
Can you show that the collection Act(G, X) of (left) actions of G on X
with your notion of morphism is a category?
2 Group representations
Now we make our first reformulation, expressing group actions as group
representations.
11
Definition 2.1. Let G be a group. A G-representation on a set X is a
group homomorphism ρ : G → AutSet (X).
Definition 2.2. Let G be a group and let ρ, σ be G-representations on
sets X and Y respectively. A morphism of G-representations f : ρ → σ is a
function f : X → Y such that
for all g ∈ G.
The condition here is best understood as the commuting of the following
diagram:
ρ(g)
X X
f f
Y Y
σ(g)
12
(a) Note that every left action of G on X is, by definition, given by a
function α : G × X → X, so that Act(G, X) ⊆ HomSet (G × X, X).
13
and
Then
and
14
(b) For G = Sn acting on X = {1, 2, . . . , n} with σ .r = σ(r), we have that
AutSet (X) = Sn and the corresponding representation ρ : Sn → Sn is
the identity map. This is the natural representation of Sn .
Furthermore, any subgroup H of Sn gives rise to a representation
ρ : H ,→ Sn in the same way. Since any finite group G is isomorphic to
a subgroup of SG (this is Cayley’s theorem, B.82), every finite group
has a permutation representation ρ : G ,→ SG .
(d) Exercise.
(e) Exercise.
3 Linear representations
As the first example above might suggest, AutSet (X) can be unwieldy to
work with. Also, for many group actions, there is additional structure
around that is ignored by looking at just the underlying sets and functions
on them. By restricting the class of representations we care about, often we
are able to say more.
With this in mind, from this point onward we will concentrate on linear
representations. That is, we will require the set G acts on to be a vector space
def
V over some field K and ask that representations take values in GL(V ) =
AutVectK (V ) (by definition of the latter).
Example 3.3. The key first example of a linear representation arises from
idGL2 (R) : GL2 (R) → GL2 (R) with idGL2 (R) (T ) = T (T : R2 → R2 being a
linear transformation, sending v to T v).
15
Indeed, if H is any subgroup of GL2 (R) then ρ : H ,→ GL2 (R) (given by
the inclusion) is a representation of H, which we call the natural represent-
ation. Clearly neither 2 nor R are special: any subgroup H of GL(V ) has a
natural representation on V , ρ : H ,→ GL(V ).
Other groups do not come to us as groups of linear transformations (or
equivalently matrices), though. So linear representation theory is precisely
about finding out in what ways the elements of our given group can be
represented by linear transformations (matrices), in a way that is compatible
with the group structure.
Example 3.4. For any group G and any vector space V , there is a repres-
entation given by ρ : G → GL(V ), ρ(g) = I for all g ∈ G, where I : V → V
is the identity linear transformation. This is the trivial representation of G
on V .
The trivial representation of degree 1 is an important special case: this
is defined by 1G : G → GL1 (K) = K \ {0}, 1G (g) = 1K for all g ∈ G.
Example 3.5. Let G = C2 be cyclic of order 2, generated by g, with g 2 = e.
Let K = C and V = C2 . Define ρ : G → GL2 (C) by
ρ(e) = ( 10 01 )
and
−5 −12
ρ(g) = 2 5
16
Exercise 3.7. Show that the dihedral group of order 8,
D8 = a, b | a4 = b2 = e, bab = a−1
0 −1
1 0
representation ρ : D8 → GL2 (Q) with ρ(a) =
has a Q-linear 1 0 and
ρ(b) = 0 −1 .
An important property of a representation is that of being faithful.
Definition 3.9. Let G be a group. Let K[G] be the K-vector space5 with
basis G. Let Lg ∈ GL(K[G]) be the linear map defined on the basis of K[G]
by Lg (h) = gh.
The regular representation of G is ρreg : G → GL(K[G]), ρreg (g) = Lg .
Furthermore. Ker ρreg = {g ∈ G | ρreg (g) = IK[G] } but we can see that
Lg = IK[G] if and only if g = e, so ρreg is faithful.
Example 3.10. Let G = C4 = a | a4 = e = {e, a, a2 , a3 }. Then an element
in K[C4 ] has the form αe + βa + γa2 + δa3 for α, β, γ, δ ∈ K. We know that
Le is the identity, since Le (h) = eh = h for all h ∈ G.
Now,
La (αe + βa + γa2 + δa3 ) = αa + βa2 + γa3 + δe
(since a4 = e), from which we see that La is represented by the matrix
0 0 0 1
1000 .
0100
0010
17
3.1 Equivalence of representations
∼
=
Let ϕ : V → W be an isomorphism of K-vector spaces. Since GL(V ) encodes
the (linear) symmetries of V , we would expect a close relationship with
GL(W ) and indeed, these are isomorphic groups. For defining
and
g : GL(W ) → GL(V ), g(U ) = ϕ−1 ◦ U ◦ ϕ
it is straightforward to check that these are inverse homomorphisms (exer-
cise).
Then a representation ρ : G → GL(V ) has a counterpart representation
f ◦ ρ : G → GL(W ) and vice versa, but these are not really “different”
representations—we have just changed the underlying vector space for an
isomorphic one. So let us say two such representations are equivalent.
G σ GL(W )
3.2 Subrepresentations
We can try to relax the conditions in the definition of equivalence. Let us
assume we have an injective linear map i : W ,→ V . Then we cannot simply
construct î as we did ϕ̂, since elements of V do not necessarily have a pre-
image in W (if i is not surjective). However we can insist that the elements
we care about do.
18
In order to make the following construction, we have to restrict to W
actually being a subspace of V (unless we want a mountain of notation and
complications). Since an injective map i : W ,→ V has W ∼= Im i ≤ V and
we have just dealt with isomorphism in the previous section, this is not a
major issue.
19
Definition 3.14. Let ρ : G → GL(V ), τ : G → GL(W ) be representations
of G. The direct sum ρ ⊕ τ is the representation ρ ⊕ τ : G → GL(V ⊕ W )
given by ρ ⊕ τ = δ ◦ (ρ × τ ) ◦ ∆.
That is,
(ρ⊕τ )(g) = (δ◦(ρ×τ )◦∆)(g) = δ◦(ρ×τ )(g, g) = δ(ρ(g), τ (g)) = ρ(g)⊕τ (g).
At this point, we will stop our study of representations (or at least, stop
doing it like this!).
4 Algebras
In this section, we will take a step back, to introduce some very general
notions, before specialising again to the context we are interested in.
When we looked at group actions, we passed to linear representations
by putting extra structure on the set being acted on. However we could
equally ask what would be the right thing to do if we had more than just
a group. In this section we will introduce modules over rings and algebras.
The motivation for doing so—beyond its intrinsic value—will have to wait
a short while, until we are ready to bring our two pictures together again.
Recall that a ring R = (R, +, ×) is an Abelian group equipped with a
compatible associative multiplication. The notion corresponding to a group
acting on a set for rings is that of an R-module. To emphasise the similarity,
we will again use the . symbol; as a map, . will take a pair (r, m) to r . m.
Definition 4.1. Let R be a ring and M an Abelian group. We say that the
pair (M, .) is a left R-module if . : R × M → M satisfies
20
Definition 4.2. Let R be a commutative ring. We say that A = (A, +, ×)
is an R-algebra if A is an R-module, A is a ring with respect to + and ×
and the two structures are compatible.
Z(A) = {z ∈ A | za = az ∀a ∈ A}.
Definition 4.3. Let V and W be K-vector spaces. The pair (X, ⊗) is said
to be the tensor product of V and W if
(c) (universal property) for every vector space Z and bilinear map h : V ×
W → Z, there exists a unique linear map h̃ : X → Z such that h =
h̃ ◦ ⊗.
⊗
V ×W X
h
h̃
Z
That is, every bilinear map h : V × W → Z factors through ⊗. A sig-
nificant advantage of this definition is that it immediately follows that the
21
vector space X is uniquely determined up to isomorphism. For if (X 0 , ⊗0 ) is
another pair satisfying the conditions of the definition, putting Z = X 0 we
have ⊗0 = ⊗ ˜0 ◦⊗ and conversely via the universal property for ⊗0 , ⊗ = ⊗◦⊗
˜ 0.
Hence ⊗ ˜ 0 ˜
˜ ◦ ⊗ = idX and ⊗ ◦ ⊗0 ˜ = idX 0 are inverse isomorphisms.
A significant disadvantage it that it is not immediate that such a pair
(X, ⊗) exists. This is frequently the case with universal property definitions:
one also needs to construct a model. (One suffices since all models will be
isomorphic.)
The standard model of the tensor product is given by taking the vector
space spanned by all symbols v ⊗ w with v ∈ V and w ∈ W , and then
imposing on this the relations which give ⊗ the bilinearity properties we
want. More formally let T denote the vector space spanned by the set of
symbols {v ⊗ w | v ∈ V, w ∈ W }. Let I be the subspace of T spanned by
the following elements:
(v1 + v2 ) ⊗ w − v1 ⊗ w − v2 ⊗ w
v ⊗ (w1 + w2 ) − v ⊗ w1 − v ⊗ w2
(λv) ⊗ w − λ(v ⊗ w)
v ⊗ (µw) − µ(v ⊗ w)
(a) (associativity)
m⊗id
A⊗A⊗A A⊗A
id⊗m m
A⊗A m A
22
(b) (unitarity)
A⊗A
u⊗id id⊗u
K⊗A m A⊗K
∼
= ∼
=
A
Here, the maps marked “∼ =” are the canonical maps sending λ ⊗ a and a ⊗ λ
to λa.
The map m is the linear map encoding multiplication, ×, and u is a map
encoding the existence of a multiplicative identity 1A (by “importing” 1K
into A).
We also have maps between algebras.
f ⊗f
A⊗A B⊗B
mA mB
A f
B
23
We can define equivalence of representations, subrepresentations and
direct sums completely analogously to the group case. However we take a
slightly different route at this point, heading towards our main theoretical
results, tying all our definitions together.
Since a K-algebra A = (A, m, u) is a ring, we can already talk about
A-modules. So, a left A-module (M, .) is an Abelian group M with a map
. : A × M → M satisfying the conditions of Definition 4.1. Now, since we
require u(1K ) = 1A to satisfy 1A . m = m, we may extend this by defining
λm = u(λ).m and hence make M a K-vector space. The additivity required
of . means that it becomes a K-linear map . : A ⊗ M → M .
Thus, in a similar spirit to above, we may give an equivalent definition
of an A-module as (M, .) with M a K-vector space and . : A ⊗ M → M
being such that the following diagrams commute:
m⊗id
A⊗A⊗M A⊗M
id⊗. .
A⊗M . M
u⊗id
K⊗M A⊗M
∼
= .
M
id⊗f
A⊗M A⊗N
.M .N
M f
N
24
Then A-Mod is a category; indeed A-Mod is a subcategory of K-Mod,
which we know by another name.
(c) The constructions in (a) and (b) are mutually inverse to each other.
Proof. This uses a much more general technical result, known as Hom-⊗
adjunction. In the case we are interested in, this says that the function
given by
def
. 7→ ρ. = (a 7→ La )
is an isomorphism, where La : M → M is the map La (m) = a . m. (We have
used the same name, La , as for the “left multiplication by a” map we have
seen for groups before; this is appropriate because the former generalises the
latter.) Note too that we have written the “long” name, HomVectK (M, M ),
for EndK (M ), so that elements of the right-hand side are indeed linear maps
A → EndK (M ), i.e. could be representations, if they have the right proper-
ties.
Elements of the left-hand side are linear maps A ⊗ M → M : we claim
that if . : A ⊗ M → M gives an A-module structure, then ρ. will be actually
be a representation.
25
To check this, we track through the definitions. For example, ρ. (ab) =
Lab so
ρ. (ab)(m) = Lab (m) = (ab).m = a.(b.m) = (La ◦Lb )(m) = (ρ. (a)◦ρ. (b))(m)
as required.
For the converse direction, let ρ ∈ HomVectK (A, HomVectK (M, M )). Then
define .ρ : A ⊗ M → M by .ρ (a ⊗ m) = ρ(a)(m). Here, ρ(a) is a map
M → M , so we can apply this to m to get an element of M . Similarly to
above,
and
G : Rep(A) → A-Mod, Gρ = (M, ϕ−1 (ρ)), Gf = f
where ρ : A → EndK (M ).
Proposition 4.12 asserts that F G = IRep(A) and GF = IA-Mod on objects
and this is clearly the case on morphisms.
26
check that ρ.N ◦ f = f ◦ ρ.M (a):
(ρ.N (a) ◦ f )(m) = ρ.N (a)(f (m))
a .N f (m)
= .N (a ⊗ f (m))
= (.N ◦ (id ⊗ f ))(a ⊗ m)
(∗)
= (f ◦ .M )(a ⊗ m)
= f (a .M m)
= f (ρ.M (a)(m))
= (f ◦ ρ.M (a))(m)
as required. The check for Gf = f (showing that the homomorphism of
representations condition translates to the homomorphism of modules con-
dition) is achieved by breaking the sequence of equalities at (∗) and noting
that then the assumption is the equality of the two ends. That is, one turns
the proof inside out!
In fact we have shown isomorphism of the categories—it is unusual one
can say something so strong.
This is the mathematical sense in which representations of A are the
same as A-modules—but even more, homomorphisms of representations are
the same as homomorphisms of modules.
To bring this back to the representation theory of groups, we next show
that there is an algebra whose representations correspond to linear repres-
entations of the group, and vice versa.
27
Proposition 5.2. There is a functor −× : AlgK → Grp given on objects
by −× (A) = A× and on morphisms by −× (f ) = f |A× for f : A → B.
Proof. We first check that −× is well-defined: this follows since if f : A → B
and a ∈ A× , f (a) ∈ B × because f is an algebra map. In more detail, we
may apply f to ab = 1A = ba to obtain
f (a)f (b) = f (ab) = f (1A ) = 1B = f (1A ) = f (ba) = f (b)f (a)
as required.
Clearly −× (idA ) = idA |A× = idA× so it remains to check composition.
Let f : A → B and g : B → C. Since (g ◦ f )|A× = g|B × ◦ f |A× , −× is a
functor.
To produce an algebra from a group, we first think about how we would
produce a vector space from a set. Well, we know how to do this: take the
K-linear span of the set (“freely”). Then the set we started with is (by con-
struction) a basis for the vector space. We could call this the “linearisation”
of the set.
Definition 5.3. Let K be a field and S a set. The linearisation of S, K[S],
is the K-vector space of K-linear combinations of elements of S.
If we wanted to be fancy (and why not?), we could go further and talk
about a functor K[−] : Set → BVectK from sets to the category of based
vector spaces whose objects are pairs (V, B) of a vector space V and a basis
B for V . The reason for using based vector spaces is that there is then a
natural forgetful functor b : BVectK → Set with b(V, B) = B, which has a
special relationship with K[−], namely forming an adjoint pair.
To do this carefully we would need to check how functions transform
under K[−], but instead we just note that the fact that linear maps are
determined by their values on a basis is precisely what is needed; we need
this fact shortly so let us give the construction a name.
Definition 5.4. Let f : S → T be a function. We define f lin : K[S] → K[T ]
to be the unique K-linear extension of f .
The next step is to show that if G is a group and not just a set, then K[G]
can be made into an algebra. We only have one bit of unused information—
the group operation, which we know is associative—so to do something
natural and general, we will have to use this to define multiplication. (We
said K[G] is a vector space so it has that as addition.)
Proposition 5.5. Let G = (G, ·) be a group. Then the maps
mG : K[G] ⊗ K[G] → K[G], mG (g ⊗ h) = g · h
and
uG : K → K[G], uG (λ) = λeG
give K[G] the structure of a K-algebra.
28
Notice that we have been lazy in defining mG , only giving its values on
a basis. As we will want this later, let us write the P most general version.
First, recall that elements of K[G] have the form g∈G αg g with αg ∈ K
(where if G is infinite, we insist that all but finitely many of the αg are zero).
We then have
X X X X X
mG ( αg g) ⊗ ( βh h) = αg βh (g · h) =
α g βh
k
g∈G h∈G g,h∈G k∈G g,h∈G
g·h=k
However, working on the basis will suffice for this proof, as long as we are
happy to accept that multiplication on the basis can be extended to linear
combinations in such a way that distributivity holds (it can).
Proof. Associativity follows from associativity in G:
(mG ◦ (mG ⊗ id))(g ⊗ h ⊗ k) = mG ((g · h) ⊗ k)
= (g · h) · k
= g · (h · k)
= mG (g ⊗ (h · k))
= (mG ◦ (id ⊗ mG ))(g ⊗ h ⊗ k).
For the unitarity,
(mG ◦ (uG ⊗ id))(λ ⊗ g) = mG (λeG ⊗ g)
= λmG (eG ⊗ g)
= λ(eG · g)
= λg
and similarly for mG ◦ (id ⊗ uG ).
Definition. Let G be a group and K a field. Then K[G] = (K[G], mG , uG )
is called the group algebra of G over K.
Proposition 5.6. There is a functor K[−] : Grp → AlgK given on objects
by K[−](G) = K[G] and on morphisms by K[−](f ) = f lin .
Proof. We need to show that if f : G → H is a group homomorphism, then
f lin : K[G] → K[H] is an algebra homomorphism: we have linearity by con-
struction and
f lin (gh) = f (g ·G h) = f (g) ·H f (h) = f lin (g)f lin (h).
Clearly idlin
G = idK[G] since
idlin
G (g) = idG (g) = g = idK[G] (g)
29
Now we have the results we have been aiming at.
Theorem 5.7. The functors K[−] : Set AlgK : −× form an adjoint pair.
Proof. The claim is that for any G ∈ Grp and A ∈ AlgK , we have a bijection
of sets
αG,A : HomAlgK (K[G], A) → HomGrp (G, A× )
that is functorial in G and A, i.e.
(a) for f : G → H,
αG,A
HomAlgK (K[G], A) HomGrp (G, A× )
−◦K[f ] −◦f
(b) for p : A → B,
αG,A
HomAlgK (K[G], A) HomGrp (G, A× )
p◦− p× ◦−
30
Now, if h : G → A× , we have K[h] : K[G] → K[A× ]. But K[A× ] ≤ A
since A× ⊆ A and A is a vector space, so K[A× ] = spanK (A× ) ≤ A (as
vector spaces). Let ιA denote the corresponding map ιA : K[A× ] → A. We
claim that αG,A (ιA ◦ K[h]) = h: we have that for all g ∈ G,
Then one may check that αG,A and βG,A are inverse to each other.
For (a), we compute: for all g ∈ G, f : G → H and h ∈ HomAlgK (K[H], A),
and
and
as required.
31
The left-hand side here is (by definition) RepK (K[G], V ) and the right-
hand side is (again by definition) RepK (G, V ). The above bijection between
these exists for any V and is natural, i.e. if V and W are two different vector
spaces, the corresponding bijections are compatible with homomorphisms
between V and W . So these bijections “fit together” to give an isomorphism
of categories between RepK (K[G]) and RepK (G).
Since we have also seen that, as categories, RepK (K[G]) ∼ = K[G]-Mod,
we have
Linear representations of G correspond to K[G]-modules.
This motivates our next chapter, studying modules and their structure.
We will begin by doing this for general algebras A and then prove some
important results that tell us that special things happen when A = K[G] for
G a finite group.
Example 5.9. Let G be a group acting on a set Ω via .. Then we can linearize
the action to a linear representation: let K[Ω] be the K-vector space with
basis Ω and define ρ : G → GL(K[Ω]) by ρ(g)(x) = g . x, extended linearly,
i.e. ρ(g) = (g . −)lin .
This representation corresponds to the K[G]-module (K[Ω], ρlin ). Fol-
lowing through the definitions, we see that the module structure map here
is really just the linearization of the original action, ..
This module is known as the permutation module associated to the action
of G on Ω.
Any algebra has a centre,
Z(A) = {z ∈ A | za = az ∀ a ∈ A},
Proof. We first show that the class sums are central. Fixing g ∈ Ci , we
may choose elements yj ∈ G (1 ≤ j ≤ m = |Ci |) so that we may write
32
Ci = {yj−1 gyj | 1 ≤ j ≤ m}, and hence C i = m −1
P
j=1 yj gyj . Then for all
h ∈ G,
Xm Xm
−1 −1 −1
h Cih = h yj gyj h = (yj h)−1 g(yj h) = C i
j=1 j=1
6 Modules
In this section, A = (A, mA , uA ) will be a K-algebra.
Recall from Section 4.2 that M = (M, .) is a left A-module if M is a
K-vector space and . : A ⊗ M → M is such that . ◦ (mA ⊗ id) = . ◦ (id ⊗ .)
and . ◦ (uA ⊗ id) = s.
We will take a categorical approach and express as many concepts in
terms of homomorphisms of modules as we can. So recall too that if (M, .M )
and (N, .N ) are A-modules, f : M → N is a homomorphism of modules if
f is a K-linear map such that f ◦ .M = .N ◦ (id ⊗ f ). An isomorphism of
modules is a bijective homomorphism.
We start with the notion of a submodule.
Definition 6.1. Let M = (M, .) be an A-module. Let N be a subspace
of M such that .(A ⊗ N ) ≤ N (that is, for all a ∈ A, n ∈ N , we have
a . n ∈ N ). Then (N, .|A⊗N ) is called a submodule of M .
We will mildly abuse notation by writing . for .|A⊗N .
One should check that (N, .) is an A-module, i.e. that the required prop-
erties for .|A⊗N follow from them being satisfied by .; essentially, we can
just observe that the conditions can be checked element-wise.
Example 6.2. The zero module is the module ({0}, . = 0) with a . 0 = 0
for all a ∈ A. We will usually save ourselves some writing by calling this
module 0 rather than {0}. Any other (sub)module is called non-zero.
33
Every non-zero module M has at least two submodules, 0 and M itself.
(The zero module has just one, of course.) Any submodule of M not equal
to M is called proper.
Example 6.3. The trivial module (not to be confused with the zero module!)
is the module (K, τ ) with τ : A ⊗ K → K, τ (a ⊗ λ) = λ for all a ∈ A, λ ∈ K.
Exercise 6.4. Check that the above definition of the trivial module is indeed
an A-module structure.
Note that under the correspondence between representations and mod-
ules, subrepresentations correspond to submodules.
This definition is “not very categorical”: if N 0 ∼
= N , N being a submod-
ule of M does not imply that N 0 is. The reason is that N is actually a subset
of M , whereas we only know that N 0 is in bijection with one. Sometimes
we really care about actual subobjects but within A-Mod, we should relax
this and consider morphisms that are monic. In A-Mod, this is equivalent
to being an injective homomorphism.
Notice that we do not need to ask for anything extra on N —just that it
is a submodule—unlike groups or rings, where we need a normal subgroup
and an ideal respectively. This is mainly due to the fact that any subgroup
of an Abelian group is normal, and so quotients are defined for any subspace
of a vector space.
We show the first module property for .̄:
34
As with submodules, being a quotient is not invariant under isomorph-
ism. This time, the “right” generalisation is to consider epis, which in
A-Mod is equivalent to being a surjective homomorphism.
Lemma 6.7. Let π : M → N be a surjective A-module homomorphism.
Then N is isomorphic to a quotient module of M , namely M/ Ker π.
Conversely, if M/N is a quotient module of M by a submodule N , there
is a surjective A-module homomorphism π : M → M/N given by π(m) =
m + N.
Since Ker π is a submodule of M (exercise), there is an associated in-
jective homomorphism ι so that we could write
ι π
Ker π M N
At M , we have π ◦ ι = 0 (everything in Im ι = Ker π is sent to 0 by π). Now
ι being injective means that if we write 0 : 0 → Ker π for the (unique) map
sending 0 to 0, we have Ker ι = 0 = Im 0. Similarly, writing 0 : N → 0 for the
(unique) map defined by 0(n) = 0 for all n ∈ N , we have Im π = N = Ker 0;
this is equivalent to π being surjective. We usually suppress the labels for
zero maps and write all of this as
ι π
0 Ker π M N 0
noting that at each object X, Im( → X) = Ker(X → ). We call this
equality of the image of the incoming map with the kernel of the outgoing
map exactness at X. Replacing Ker π by anything isomorphic to it, we
obtain the following definition.
Definition 6.8. Let K, M and N be A-modules.
If f : K → M and g : M → N are module homomorphisms such that f
is injective, g is surjective and Im f = Ker g, then we say that
0 K M N 0
is a short exact sequence of A-modules.
It is called “short” because, well, there are long ones too. Short exact
sequences are sometimes called extensions, because they describe a way to
“extend” K by N to form a bigger module M , having a submodule iso-
morphic to K with quotient isomorphic to N .
Some short exact sequences are special: this is when we can actually find
a submodule of M isomorphic to N too.
Definition 6.9. A short exact sequence
f g
0 K M N 0
of A-modules is split if there exists an A-module homomorphism u : N → M
such that g ◦ u = idN .
35
Note that u is necessarily injective. In nice categories—the correct name
is Abelian categories—such as A-Mod, split short exact sequences arise in a
particularly natural way.
Definition 6.10. Let M = (M, .M ) and N = (N, .N ) be A-modules. Then
the direct sum of vector spaces M ⊕ N is given an A-module structure via
.⊕ : A ⊗ (M ⊕ N ) → M ⊕ N, .⊕ (a ⊗ (m + n)) = a .M m + a .N n
Then show that setting u = h−1 ◦ ιN we have a map showing that the
sequence splits.
This is the starting point for the construction of an object that classi-
fies extensions of modules. The details would require too much time and
technicalities beyond us at this point, but the idea is accessible: take all
short exact sequences starting with K and ending with N , define an equi-
valence relation on these and hence construct a group Ext1 (K, N ), in which
[0 → K → K ⊕ N → N → 0] is the identity element. In particular,
Ext1 (K, N ) = {0} if and only if every extension of K by N is split (i.e. the
only extension is the trivial one, K ⊕ N ).
36
In some sense, one major goal of representation theory is to understand
Hom(K, N ), Ext1 (K, N ) and the higher extension groups Exti (K, N ) for all
modules K, N for some given algebra. In fact, there is usually a subsidiary
goal before this, namely to understand all of the “fundamental pieces” from
which we can build other modules via extensions. This is what we now turn
our attention to.
Remark 6.13 (Important!). Simple modules are also called irreducible7 , al-
though this term tends to be used more for the corresponding representation.
That is, we say a representation is irreducible if it has only the zero subrep-
resentation and itself as subrepresentations. Then it is common to shorten
“irreducible representation” to “irrep”. As we are taking a module-theoretic
approach, we will say “simple module”.
We can glue simples together:
This notion is also called “completely reducible” (usually when the term
irreducible is being used, rather than simple). This deserves a name because
in general not every module is semisimple—this is not meant to be obvious,
but it is (very much) true.
Proof. We will omit this proof, as in full generality (to deal with the “not
necessarily finite” cases) it requires some more advanced techniques. Those
who are interested will find this result in [Rot, Section 4.1] or [EH, Sec-
tion 4.1].
37
Definition 6.16. Let M be an A-module. A composition series for M is a
sequence of submodules
0 = M0 ≤ M1 ≤ M2 ≤ · · · ≤ Mr−1 ≤ Mr = M
such that Mi+1 /Mi is simple for all 0 ≤ i ≤ r − 1. We call r the length of
the composition series.
0 ≤ M1 ≤ · · · ≤ Mr−1 ≤ Mr = M
0 ≤ N1 ≤ · · · ≤ Ns−1 ≤ Ns = M
Since then any two composition series for a module M have the same
length, we may say that M has (well-defined) length, the length of any
composition series for it.
Now, if M has a composition series
0 ≤ M1 ≤ · · · ≤ Mr−1 ≤ Mr = M
we have
• M1 is simple;
38
• 0 → M2 → M3 → M3 /M2 → 0 is a short exact sequence with M3 /M2
also simple;
• ...
39
Since λidS ∈ EndA (S) for all λ ∈ K, {λidS | λ ∈ K} ⊆ EndA (S).
Conversely we have shown above that EndA (S) ⊆ {λidS | λ ∈ K}. Hence
these vector spaces are equal; note finally that {λidS | λ ∈ K} ∼
= K as
algebras.
Our primary goal shortly will therefore be to classify the simple K[G]-
modules.
Before that, though, we take what appears to be a diversion via another
approach to “building blocks”.
M = M1 ⊕ M2 ⊕ · · · ⊕ Mr = N1 ⊕ N2 ⊕ · · · ⊕ Ns
40
Definition 6.25. Let M be a left A-module. We say M is a finitely gener-
ated free module if M ∼
= A A⊕n for some n ∈ N.
Dually,
(a) P is projective;
(a) I is injective;
41
Proposition 6.30. The following are equivalent:
Proof. We first prove (a) if and only if (d), then (b) if and only if (c) and
finally (a) if and only if (b).
• (a)⇐⇒(d)
Assume (a). Then every free module A A⊕I is also semisimple. Since
every module M ∈ A-Mod is a quotient of a free module (at worst, we
may take I to be an indexing set for a basis of M ) and quotients of
semisimple modules are semisimple, we have (d).
Conversely (d) implies (a), by definition.
• (b)⇐⇒(c)
Assume (b) and let f : M P be a surjection. Then Ker f ≤ M is
injective so M = Ker f ⊕ Q for some Q. But by the first isomorphism
theorem, Im f = P ∼ = Q. Then this isomorphism splits f and P is
projective.
Assume (c) and let f : I ,→ M be an injection. Then Coker f = M/I
is projective so π : M → M/I splits, via r : M/I → M , and M =
Ker π ⊕ Im r. But Ker π = I so I has a complement in M and is
injective.
• (a)⇐⇒(b)
Assume (a). Then as above, any M ∈ A-Mod is a quotient of a free
module A A⊕I via π : A A⊕I M and this free module is semisimple.
Hence A A⊕I = Ker π ⊕ Q with Q ∼
= M , so M is projective.
Assume (b). Let M be a submodule of A A. Then since A A/M is
projective, π : A A → A A/M splits and M has a direct complement.
By Lemma 6.15, A A is semisimple.
Proof.
42
(a) Let S ∈ A-Mod be a simple A-module and let s ∈ S \ {0}. Then there
is an A-module homomorphism f : A A → S given by f (a) = a .S s for
all a ∈ A; f is linear by the properties of actions and a .S (f (b)) =
a .S (b .S s) = (ab) .S s = f (ab). Now since S is simple and s 6= 0,
Im f = S and f is surjective.
Then since A semisimple implies S is projective, f splits, i.e. S is
(isomorphic to) a direct summand of A A.
L
(b) We have that A A = i∈I Si . Now 1A ∈ A A so there exists a finite
subset
P J ⊆ I and elements sj ∈ Sj for all j ∈ J such that 1A =
j∈J sj .
P L
But then for all a ∈ A, a = a1A = a s
j∈J j ∈ Sj . So
L L L L j∈J
j∈J Sj ⊆ i∈I Si = A A ⊆ j∈J Sj and hence A A = j∈J Sj .
43
For any field K, there is a (non-zero) ring homomorphism u : Z → K,
u(1Z ) = 1K (extended linearly: u(n1Z ) = n1K ). The kernel of Ker u is a
proper ideal of Z, therefore10 either Ker u = 0 or Ker u = nZ for some
n ∈ Z.
Now since K is a field, Im u is an integral domain and Im u ∼
= Z/nZ by
the First Isomorphism Theorem (B.144). But Z/nZ is an integral domain
if and only if n is prime. So Ker u = 0 or Ker u = pZ for some prime p.
If Ker u = 0, we say K has characteristic zero; if Ker u = pZ with p
prime, we say K has characteristic p. We denote the characteristic of K by
char K.
Theorem 6.34 (Maschke). Let G be a finite group. Then the group algebra
K[G] is semisimple if and only if char K does not divide |G|.
Note that 0 always divides |G|, i.e. over fields of characteristic zero—such
as R and C—the group algebra of a finite group is always semisimple.
Before we prove Maschke’s theorem, we need a lemma.
Lemma 6.35. Let G be a finite group and K a field. Let M and N be
K[G]-modules and let f : M → N be a K-linear map. Then
X
T (f ) : M → N, T (f )(m) = g . (f (g −1 . m))
g∈G
is a K[G]-module homomorphism.
Proof. Exercise.
44
def P
Now assume K[G] is semisimple. Consider w = g∈G g ∈ K[G]. Then
hw = w for all h ∈ G so Kw ≤ K[G] K[G] is a 1-dimensional submodule of
K[G] K[G]. Since K[G] is semisimple, there exists C such that as modules we
have K[G] K[G] ∼= Kw ⊕ C.
Then there exists c ∈ C such that 1K[G] = λw + c for λ ∈ K, c ∈ C. We
have λ 6= 0, else K[G] K[G] = K[G]C ⊆ C 6= K[G] K[G], a contradiction.
We see that w2 = |G|w and
Remark 6.36. With a little more technology, one can prove the stronger
form of Maschke’s theorem: K[G] is semisimple if and only if G is finite and
char K divides |G|.
Consequently,
Pr 2
(c) |G| = i=1 ni
Proof.
45
(b) The matrix algebra Mni (K) has exactly one simple module, the natural
module, isomorphic to K⊕ni , of dimension niQ . Set Si = K⊕ni . It is
straightforward to check that all simples for ri=1 Mni (K) are of the
form
0 × 0 × · · · × 0 × Si × 0 × · · · × 0.
Furthermore Mni (K) is a simple algebra of dimension n2i so
ei 1K[G] = ei (e1 + · · · + er ) = ei e1 + · · · + ei er
46
If all simples are 1-dimensional, then ni = 1 for all i so K[G] ∼
= K⊕r is
commutative, hence G is Abelian. Then by (c),
r
X r
X
|G| = n2i = 12 = r.
i=1 i=1
The take-away from this theorem is that for finite groups and over algeb-
raically closed fields of characteristic zero, we have very detailed information
on their representation theory. Several of these details will re-emerge shortly,
when we look at characters.
First, let us specialise one step further and look at what this means for
the simplest groups, namely finite Abelian groups. To avoid field-theoretic
complications, let us take K = C.
If G is a finite Abelian group, then
G∼
= Cn1 × Cn2 × · · · × Cnr
where Cni = hgi | gini = ei. In particular, if g ∈ G we can write (by a mild
abuse of notation) g = g1a1 · · · grar with ai ∈ Z.
Then letting ζi ∈ C be an ni th root of unity and ζ = (ζ1 , . . . , ζr ), define
Vζ to be the 1-dimensional module Vζ = Cv with
7 Character theory
From this point, unless otherwise specified, G will be a finite group, all
vector spaces will be C-vector spaces and all C[G]-modules will be finite-
dimensional left C[G]-modules.
A complex matrix representation of a finite group ρ : G → GLn (C) con-
tains a large amount of data: |G|n2 complex numbers, in fact. Passing from
a representation to its character discards the majority of this but retains
enough to be useful, as we will see. The character of ρ is the map χ : G → C
given by tr ◦ ρ, taking each group element to the trace of the representing
matrix, and so keeps only |G| pieces of data. One may alternatively prefer
to define the character of a module:
47
Definition 7.1. Let M = (M, .) be a C[G]-module with basis B. The
character of M is the function χM : G → C defined by χM (g) = tr(MBB (Lg )),
where Lg : M → M is the endomorphism of M given by Lg (m) = g . m.
Examples 7.3.
and this character is called the regular character. (This formula for
χreg is a consequence of (b) with X = G, g . h = gh, as no element
other than e fixes an element of G: g . h = gh = h ⇒ g = e.)
48
Proposition 7.5. Let χ be a character of G and thus the character of some
C[G]-module M . Then
(a) χ(e) = dim M ;
tr(MBB (Lh−1 gh )) = tr(MBB (Lh )−1 MBB (Lg )MBB (Lh )) = tr(MBB (Lg ))
(c) Since G is finite, g ∈ G has some finite order m. Then M = MBB (Lg )
satisfies xm − 1, which splits as distinct linear factors over C, hence
the minimum polynomial of M does and so M is diagonalisable. This
implies that χ(g) = tr(M) is the sum of the eigenvalues of M and
since these are roots of xm − 1, they are mth roots of unity.
Now the eigenvalues of M−1 = MBB (Lg−1 ) are inverses of roots of unity,
but if ω is a root of unity then ω −1 = ω and hence χ(g −1 ) = χ(g).
The proof of (c) also tells us that |χ(g)| ≤ χ(e) = n and if g has order 2
then χ(g) ∈ Z and χ(g) ≡ χ(e) (mod 2) (exercise).
Definition 7.6. The degree of a character χ of G is χ(e). Characters of
degree 1 are called linear characters.
The trivial character is linear; the regular character has degree |G|. The
map sign : Sn → C× given by taking the sign of a permutation is a linear
character of Sn .
Lemma 7.7. Let G be a group and χ : G → C a character of G. The
×
codomain restriction χ|C : G → C× is a group homomorphism if and only
if χ is a linear character.
Proof. We sketch the implication “linear ⇒ homomorphism”. Since χ is
linear, i.e. is the character of a module M of dimension 1, we may identify
M with C and take the basis B = {1} for it.
Then for any g ∈ G, g . 1 = λg ∈ C. Then Lg is the map λg idC , or
equivalently MBB (Lg ) = (λg ), a 1 × 1 matrix. Now tr((λg )) = λg , from which
×
it is straightforward to check that χ|C is a homomorphism.
We leave the converse as an exercise.
49
Remark 7.8. Let ρ : G → GLn (C) be a representation of G and χ : G → C
its character. Then |χ(g)| = χ(e) for all g ∈ G if and only if ρ(g) = λIn for
some λ ∈ C. Hence the kernel of ρ is equal to the set {g ∈ G | χ(g) = χ(e)},
so that the character retains full information on the kernel of its originating
def
representation. We then set Ker χ = {g ∈ G | χ(g) = χ(e)} = Ker ρ.
e g g2 g3
χ1 1 1 1 1
χ−1 1 −1 1 −1
χi 1 i −1 −i
χ−i 1 −i −1 i
50
Since for characters, ψ(g) = ψ(g −1 ) and {g −1 | g ∈ G} = G, it is easy
to see that when we restrict this inner product to the set of characters of
G, we have hϕ, ψi = hψ, ϕi, so hϕ, ψi is in fact real and h , i is symmetric,
rather than just conjugate-symmetric.
Recall that we may write C[G] C[G] = ri=1 Bi where each pair of sub-
L
⊕n
modules Bi ∼ = Si i and Bj ∼
⊕n
= Sj j have no common composition factor.
Recall also that there exist {ei | 1 ≤ i ≤ r} a complete set of orthogonal
central idempotents, with each ei spanning Z(Bi ).
Hence hχi , χi i = 1. L
For the case i 6= j, set Y = Bi ⊕ Bj and Z = k6=i,j Bk , so Y and Z
have no common composition factors and C[G] C[G] = Y ⊕ Z. Now letting
{ei } be as above, we see from the proof of the previous proposition that
1 X
ϕi (g −1 ) + ϕj (g −1 ) g
ei + ej =
|G|
g∈G
51
and the coefficient of e in ei + ej is (ϕi (e) + ϕj (e))/|G|. Denote by λg the
coefficient of g in |G|(ei + ej ), so λg = ϕi (g −1 ) + ϕj (g −1 ). Then
2
1 X 1 XX
(ei + ej )2 = 2
λg g = (λh λh−1 g )g
|G| |G|2
g∈G g∈G h∈G
Also
1 X
ϕi (h) + ϕj (h) ϕi (h−1 ) + ϕj (h−1 )
hϕi + ϕj , ϕi + ϕj i =
|G|
h∈G
Theorem 7.15. Two C[G]-modules are isomorphic if and only if their char-
acters are equal.
52
Proof. We discussed the forward direction earlier.
If V , W are finite-dimensional C[G]-modules with equal character χ then
⊕β
∼
V = ri=1 Si⊕αi , W ∼ = i=1 Si i and αi = hχ, χi i = βi , so V ∼
L Lr
= W.
Proof. This follows straightforwardly from C[G] being semisimple (and hence
all its modules being semisimple), orthogonality and Schur’s lemma, which
tells us that dim HomC[G]-Mod (Si , Sj ) = δij .
Proof. The first claim is simply a restatement of Theorem 7.13, making use
of the fact that characters are class functions and that |Ci ||CG (gi )| = |G|.
For the second part, let {δi | 1 ≤ i ≤ r} be the aforementioned basis for
ClassFun(G, C), with δi taking value Pr1 on jCi and 0 elsewhere.
j
Since Irr(G)
is a basis for ClassFun(G, C), δj = k=1 λk χk for some λk ∈ C. Then
1 X 1 χk (gj )
λjk = hδj , χk i = δj (g)χk (g) = |Cj |δj (gj )χk (gj ) = .
|G| |G| |CG (gj )|
g∈G
53
Note that the character table is an invertible matrix: by orthogonality,
its rows—the irreducible characters—are linearly independent. Also since
r
X r
X
2
χi (e) = (dim Si )2 = |G|
i=1 i=1
we see that the sum of the squares of the entries in the first column equals
|G|. We will refer to this as “the degree sum property”.
Examples 7.20.
(a) G = C2 = a | a2 = e ; C1 = {e}, C2 = {a} so set g1 = e, g2 = a.
gi e a
|CG (gi )| 2 2
χ1 1 1
χ2 1 −1
• χ1 is the trivial character
• χ2 (e) = 1 since |G| = 2 = 12 + 12
• χ2 (a) = −1 by row orthogonality: 1 · 1 + 1 · χ2 (a) = 0 implies
χ2 (a) = −1
Note that χreg = χ1 + χ2 .
(b) G = S3
gi ι (1 2 3) (1 2)
|CG (gi )| 6 3 2
χ1 1 1 1
χ2 1 1 −1
χ3 2 −1 0
• χ1 is the trivial character; χ2 is the sign character
• χ3 (ι) is obtained from the degree sum property: 22 = 6 − 12 − 12
• χ3 ((1 2 3)) is obtained by column orthogonality:
54
(c) G = C2 × C2 = a, b | a2 = e, b2 = e, ab = ba
gi e a b ab
|CG (gi )| 4 4 4 4
χ1 1 1 1 1
χ2 1 1 −1 −1
χ3 1 −1 1 −1
χ4 1 −1 −1 1
Here χreg = χ1 + χ2 + χ3 + χ4 .
In order to further aid our search for characters of a group, we return to
linear characters; that is, characters of degree 1.
Proposition 7.21. Let λ be a linear character of G and χ any character.
Then the product χλ : G → C defined by (χλ)(g) = χ(g)λ(g) for all g ∈ G
is a character of G. Furthermore, χλ is irreducible if and only if χ is
irreducible.
Proof. Exercise.
55
and thus character χ. Then χ(e) = χ e(eN ) = χ e(eG/N ) so the degrees are
equal.
The rest of the proof is deferred to the following theorem, which shows
that a slightly stronger statement than this proposition in fact holds.
G0 = ghg −1 h−1 | g, h ∈ G
Proof. Exercise
Proposition 7.26. If
T N P G then there exist irreducible characters χi1 , . . . , χis
of G such that N = sk=1 Ker χik .
Proof. If g ∈ Ker χ for all χ ∈ Irr(G) then g ∈ Ker δ1 (where δ1 is the class
function with value 1 on C1 = {e} and 0 otherwise), T since Irr(G) is a basis
for the space of class functions. So g = e and χ∈Irr(G) Ker χ = {e}.
eis }, so {N } = sk=1 Ker χ
T
Let Irr(G/N ) = {e χi1 , . . . , χ eik by the above. Let
χik be theT lift of χ
eik to G and note T that g ∈ Ker χik implies gN ∈ Ker χ eik .
So if gT∈ sk=1 Ker χik then gN ∈ sk=1 Ker χ eik = {N } and so g ∈ N . Hence
N = sk=1 Ker χik .
56
Proof. We prove that G is not simple if and only if Ker χi 6= {e} for some
non-trivial irreducible character χi . If N is a proper non-trivial normal
subgroup of G, then by the previous proposition there exists an irreducible
character χi whose kernel is neither {e} nor G, as required. Conversely,
given a non-trivial χi with non-trivial kernel, this kernel is a proper non-
trivial normal subgroup.
We see from its character table that S3 is not simple: χ2 ((1 2 3)) = 1 =
χ2 (ι). Indeed the sign character shows that Sn is not simple for all n ≥ 3.
We return to giving some more examples, beginning with A4 .
Examples 7.28.
(a) G = A4
gi ι (1 2)(3 4) (1 2 3) (1 3 2)
|CG (gi )| 12 4 3 3
χ1 1 1 1 1
χ2 1 1 ω ω2
χ3 1 1 ω2 ω
χ4 3 −1 0 0
(b) G = S4
gi ι (1 2) (1 2 3) (1 2)(3 4) (1 2 3 4)
|CG (gi )| 24 4 3 8 4
χ1 1 1 1 1 1
χ2 1 −1 1 1 −1
χ3 2 0 −1 2 0
χ4 3 1 0 −1 −1
χ5 3 −1 0 −1 1
57
the table, since (1 2)(3 4) ∈ V4 implies χ3 ((1 2)(3 4)) = χ
e3 (ι) = 2
and V4 (1 2 3 4) = V4 (1 3) implies
χ3 ((1 2 3 4)) = χ
e3 ((1 3)) = χ
e3 ((1 2)) = 0.
gi e a a2 a3 b
|CG (gi )| 14 7 7 7 2
χ1 1 1 1 1 1
χ2 1 1 1 1 −1
χ3 2 + −1 2 + −2 3 + −3 0
χ3 2 2 + −2 4 + −4 6 + −6 0
χ3 3
2 + −3 6
+ −6 2
+ −2 0
Q = a, b | a4 , a2 = b2 , b−1 ab = a−1 ,
58
Proof. Exercise.
At this point, we will remark that one can gain information about the
representation theory of more general algebras than just group algebras by
the use of characters. However, the group setting is where it is most effective
and for arbitrary algebras, especially finite-dimensional ones, there are other
techniques.
These include quivers: on the one hand, we can associate a quiver11
to an algebra from its structure, and then use the representation theory of
quivers to study it. On the other, there are natural quivers appearing in
the category of modules, the most important being the Auslander–Reiten
quiver. This requires a whole other course in representation theory, but for
those interested, [EH], [Sch] and [ASS] are recommended reading.
11
This is another name for a directed graph.
59
A Burnside’s pα q β theorem
There are many other important and interesting constructions to help find
characters but we cannot describe them all here. Some of the most significant
include tensor products, restriction and induction and Frobenius reciprocity
but in this appendix we explain one of the most significant applications of
representation theory to structure theory: Burnside’s pα q β theorem.
To appreciate this example fully one needs some advanced group theory.
In particular, this appendix is recommended to be read alongside a course
on Galois Theory.
We will begin with some properties of algebraic integers necessary for our
proof of Burnside’s pα q β theorem and, along the way, demonstrate a pair of
very useful theorems, namely that rational character values are necessarily
integers and that the degree of any irreducible character divides the group
order.
Definition A.1. A complex number λ is said to be an algebraic integer if
it is the eigenvalue of an integer matrix.
Equivalently, λ is an algebraic integer if it is the root of some monic
polynomial in Z[x]. So every integer is an algebraic integer, as is every
√ 0 1 )). If λ is an
positive square root m of a natural number m (consider ( m 0
algebraic integer so are −λ and λ. Also every root of unity is an algebraic
integer (consider xn − 1). Indeed, if λ and µ are algebraic integers then so
are λµ and λ + µ, so the set of algebraic integers A is a subring of C. It
follows that every character value χ(g) is an algebraic integer, as we showed
earlier that character values are sums of roots of unity.
The next theorem tells us even more about character values.
Theorem A.2. Any rational algebraic integer is an integer: Q ∩ A = Z.
Proof. Suppose λ ∈ Q \ Z. Write λ = r/s with (r, s) = 1 and s 6= ±1. Let
p be a prime dividing s. Then if A ∈ Mn (Z), the off-diagonal entries of
sA − rI are divisible by s and hence by p. So det(sA − rI) = (−r)n + mp
for some m ∈ Z. Since p - r we conclude that det(sA − rI) 6= 0. Then
n
det(A − λI) = 1s det(sA − rI) 6= 0 for all A ∈ Mn (Z) and λ is not an
algebraic integer.
60
Proof. The existence of u gives us that λ is an eigenvalue of the integer
matrix A = (αg−1 gj ) where {g1 , . . . , gn } is an enumeration of the elements
i
of G.
|G| χ(g)
Lemma A.6. If χ is an irreducible character of G and g ∈ G then |CG (g)| χ(e)
is an algebraic integer.
Proof. Let S be a simple C[G]-module
P with character χ, C be the conjugacy
class containing g and C = h∈C h the class sum. We claim that uC = λu
for all u ∈ S, where λ = |C|G| χ(g)
G (g)| χ(e)
.
Since C ∈ Z(C[G]) there exists some
P λ ∈B C such that uC = λu for all
u
P ∈ V . If B is a basis for S then h∈C MB (Lh ) = λI so taking traces,
h∈C χ(h) = λχ(e). But χ is constant on conjugacy classes, hence on C,
so |C|χ(g) = λχ(e) and hence λ = |C|G| χ(g)
G (g)| χ(e)
. We conclude that λ is an
algebraic integer, by the previous lemma.
Therefore we have:
Theorem A.7. If χ is an irreducible character of G then χ(e) divides |G|.
Proof. Let {gi | 1 ≤ i ≤ r} be a complete set of conjugacy class repres-
entatives. Then for all i, |CG|G| χ(gi )
(gi )| χ(e) and χ(gi ) are algebraic integers, so
Pr |G| χ(gi )χ(gi )
i=1 |CG (gi )| χ(e) is an algebraic integer. But this algebraic integer is
equal to |G|/χ(e) by the row orthogonality relations. So the rational algeb-
raic integer |G|/χ(e) is an integer and hence χ(e) | |G|.
61
Corollary A.10. All character values for symmetric groups are integers.
Now we can turn directly towards our main goal. For this, we need an
extension of the notion of an algebraic integer, namely algebraic numbers.
A complex number is said to be algebraic (or an algebraic number) if it is
the root of some non-zero polynomial in Q[x]. An algebraic number α has
a minimum polynomial m(x) and we will refer to the elements of the set of
roots of m(x) as conjugates of α. We will need the following facts about
algebraic numbers.
62
the left-hand side is an algebraic integer and χi (g) 6= 0 so it is non-zero. So
χi (g)/χi (e) ∈ A and hence |χi (g)/χi (e)| = 1.
Let ρ be a representation of G with character χi . Then by Exercise 7.8,
there exists λ ∈ C such that ρ(g) = λI. Let K = Ker ρ C G, a proper normal
subgroup since χi 6= χ1 . If K 6= {e} then G is not simple. If K = {e} then
gρ = λI commutes with hρ for all h ∈ G and so g commutes with every
h ∈ G, since K = {e} means that ρ is faithful. So g ∈ Z(G) C G and
Z(G) 6= {e}, so G is not simple.
• α = β = 0: G = {e} is soluble
63
B Groups, rings and vector spaces
An algebraic structure is just a set with some operations. We might ask that
the operations have certain properties or that different operations interact
in a particular way. In principle there are many different types of algebraic
structure but in practice some have much richer theories than others and
more (interesting) examples “in nature”.
An algebraic structure consisting of a set S and n operations will be
denoted by (S, ∗1 , ∗2 , . . . , ∗n ). Sometimes we will use different letters for
the set S, to help us remember which type of algebraic structure we are
thinking about, and we might use established or more suggestive notation
for the operations, such as + or ×.
Definition B.1. A group is an algebraic structure (G, ∗) with ∗ an associ-
ative binary operation, having an identity and such that every element of G
has an inverse with respect to ∗.
Definition B.2. A group (G, ∗) is called Abelian if ∗ is commutative.
Definition B.3. A ring is an algebraic structure (R, +, ×) such that (R, +)
is an Abelian group, × is associative and we have
a × (b + c) = (a × b) + (a × c)
and
(a + b) × c = (a × c) + (b × c)
for all a, b, c ∈ R. The identity element for the group (R, +) will be denoted
0R .
The two equations given here are called the first and second (or left and
right) distributive laws. We could be more general and say that × distributes
over + from the left if a × (b + c) = (a × b) + (a × c), and correspondingly
for the right-handed version. Notice that these are not symmetric in + and
×.
Definition B.4. A field is an algebraic structure (F, +, ×) such that (F, +, ×)
is a ring, × is commutative and has an identity 1F 6= 0F and every element
of F \ {0F } has an inverse with respect to ×.
Notice that the definition of a field (F, +, ×) could also be written
equivalently as saying that (F, +) is an Abelian group with identity 0F ,
(F \ {0F }, ×) is an Abelian group and that + and × satisfy the distributive
laws.
From now on, we will generally prefer the symbol K for a field; this will
avoid clashes of notation as we have another concept for which we will prefer
F.
64
We see from these definitions that rings are groups with extra structure,
and fields are rings with extra properties. So every field is in particular a
ring, and every ring is a group.
Groups, rings and fields appear in various different contexts and the
following list of examples is not in any way comprehensive. Rather, they are
examples that should be somewhat familiar to you following a first course
in abstract algebra.
Number systems
• Q, R and C are fields.
• Z is a ring (but not a field); in many ways, Z is the prototype ring and
many ring-theoretic questions are inspired by the structure of Z.
• For each n, the set of integers modulo n, Zn (with addition and mul-
tiplication modulo n) is a ring.
• Since Q, R and C are fields, as we said in the remarks after the defin-
ition of a field, we have associated groups Q× = Q \ {0}, reals× =
R \ {0} and C× = C \ {0} with respect to multiplication.
More generally, the set of invertible elements in a ring R forms a group
under multiplication (called the group of units, R× ). For a field K, we
have K× = K \ {0K }, because every non-zero element is invertible.
Symmetry groups
• Natural examples of symmetry groups are the set of rotations and
reflections of a square, an n-gon, a circle or other 2-dimensional figures,
as well as rotations and reflections of a cube, tetrahedron, etc.
A particularly important example that occurs often is the dihedral
group D2n (of order 2n), defined to be the group of isometries of R2
preserving a regular n-gon. The dihedral group D2n is generated by
two elements α and β and has the following presentation:
D2n = a, b | αn = e, β 2 = e, αi β = βα−i ∀ i ∈ Z
65
• The “symmetries of a set” S are given by the set of bijections
def
Bij(S) = {f : S → S | f is a bijection}. This set satisfies all the
conditions to be a group with respect to composition of functions.
Permutations are exactly bijections from the set {1, . . . , n} to itself.
These groups are called the symmetric groups.
They (and the symmetry groups of geometric objects) are usually not
Abelian, in contrast to all the previous examples.
Matrices
• We can add m×n matrices and multiply n×n matrices, we know what
the zero and identity matrices are and we know that sometimes we can
invert matrices. So it should not be a surprise that sets of matrices
can have either group or ring structures, though they are almost never
fields.
More precisely, Mn (Z), Mn (Q), Mn (R) and Mn (C) are all rings; in fact
Mn (R) is one for any ring R. We call these matrix rings. (In linear
algebra, we usually work over a field, because we want to be able to
divide by scalars as well as multiply by them.)
• Inside any matrix ring over a field K, the subset of invertible matrices
(those with non-zero determinant) forms a group, the group of units,
as defined above. We have a special name and notation for the group
of units of matrix rings: we call them general linear groups and write
GLn (K). The group operation is matrix multiplication.
Sets of functions
• The set of functions from R to R, denoted F(R, R), is a ring with
respect to pointwise addition and multiplication, that is, for functions
f, g and x ∈ R, (f + g)(x) = f (x) + g(x) and (f g)(x) = f (x)g(x).
This list might give the (erroneous) impression that there are lot more
rings than groups. This is not the case: since every ring is a group and not
every group is a ring, there are definitely more groups than rings. It just so
happens that the examples you are familiar with are mostly rings or fields.
One important observation about the above list is that lots of these ex-
amples, notably the symmetric groups, the matrix rings and general linear
66
groups, are naturally symmetries of something. The description of sym-
metry groups we gave says this explicitly. One of the main results in Lin-
ear Algebra is that matrices are the same thing as linear transformations
of vector spaces, and hence GLn (K) is the symmetry group of Kn , the n-
dimensional vector space over K. This idea precisely leads to representation
theory.
Having mentioned vector spaces, let us briefly say where they fit in terms
of the definitions so far. Every vector space V over a field K is an Abelian
group under addition, (V, +). Vector spaces are not rings (or fields): we do
not multiply vectors by vectors.
However vector spaces have more structure than just addition, namely
they have scalar multiplication. Scalar multiplication is a slightly different
type of operation: it is not a binary operation on V . It takes two inputs, a
scalar λ ∈ K and a vector v ∈ V , and produces a new vector λv ∈ V . We
can formalise this as a function · : K × V → V , λ · v = λv, and say that a
vector space is an algebraic structure (V, +, ·) where + and · satisfy some
compatibility conditions, namely those that give linearity.
B.1 Groups
Definition B.5. A group (G, ∗) is called finite if the set G is finite.
Recall that the set of bijections from a set S to itself is a group with
respect to composition of functions, (Bij(S), ◦). Let us fix S = {1, 2, . . . , n}
for some n ∈ N.
67
Remark B.10. In fact, a little more is true: La is actually surjective, as we
may easily show since we have already done all the work needed. Let b ∈ G.
Then by part (iv), there exists a solution to the equation a ∗ x = b, namely
a−1 ∗ b.
So for any fixed a ∈ G, La : G → G is in fact a bijection, that is, it is a
permutation of the elements of G. Of course, the same is true of Ra .
Proposition B.11. Let (G, ∗) be a group and let a, b ∈ G. Then
Proposition B.12. Let (G, ∗) and (H, ◦) be groups. Then the Cartesian
product G × H is a group with respect to the binary operation • defined by
for all g1 , g2 ∈ G, h1 , h2 ∈ H.
We briefly recap the notion of the sign of a permutation.
Let π ∈ Sn and suppose that π = γ1 γ2 . . . γk with the γi disjoint cycles.
Set
k
X
ν(π) = (|γi | − 1),
i=1
where |γi | denotes the length of the cycle γi . Note that ν(π) is well-defined,
although this requires some care (in particular, cycles of length 1 contribute
0 to the sum, and so may be ignored). Note that π −1 has the same cycle
lengths as π, so we have ν(π −1 ) = ν(π).
Clearly ν(π) = 1 if and only if π is a transposition. In fact we may
interpret ν(π) as follows:Pif all possible 1-cycles are included in the expression
k
π = γ1 γ2 . . . γk , then i=1 |γi | = n; thus ν(π) = n − k, i.e. ν(π) is the
difference between n and the number of cycles in the full cycle notation.
Definition B.13. Given π ∈ Sn , the sign of π is defined by
sign(π) = (−1)ν(π) .
68
Lemma B.15. Let π ∈ Sn and let τ ∈ Sn be a transposition. Then
sign(πτ ) = (−1) · sign(π) = − sign(π).
sign(τ1 · · · τr ) = (−1)r .
0 0 1 0 0 0 0 1
69
0 0 0 1
0 1 0 0
Aστ = A(1 4) =
0
and
0 1 0
1 0 0 0
0 0 0 1 0 1 0 0 0 0 0 1
1 0 0 0 0 0 1 0 0 1 0 0
Aσ Aτ =
0
= = Aστ .
1 0 0 1 0 0 0 0 0 1 0
0 0 1 0 0 0 0 1 1 0 0 0
B.2 Subgroups
Definition B.20. Let (G, ∗) be a group. A subset H of G is said to be a
subgroup if (H, ∗) is a group. If this is the case, we write H ≤ G.
Remarks B.21.
Examples B.22.
(a) (Z, +) is a subgroup of (Q, +), and both are subgroups of (R, +).
(c) (Q∗ , ×) is not a subgroup of (Q, +), since the two binary operations
are different.
70
(e) In the group GLn (R) (where the operation is matrix multiplication),
the set of matrices with determinant 1 is a subgroup; this follows
from properties of determinants you proved in Linear Algebra. This
subgroup has a special name: it is known as the special linear group,
SLn (R). (As before, one may replace R by any field F and obtain
SLn (F ).)
This last pair of examples is “uninteresting” (in the sense that they don’t
tell us anything we didn’t already know) and on occasion we will want to
exclude these from claims about subgroups. So we introduce the following
terminology.
(SG1) eG ∈ H;
71
Be alert here: it is not enough for a subset to be closed under the group
operation for it to be a subgroup. It is possible to find examples of sub-
sets satisfying (SG2) but not one or other of (SG1) and (SG3), as we will
see shortly. But first, let us start off positively and see some examples of
subgroups and the subgroup test in action.
Examples B.25.
(a) The set of even integers 2Z = {2r | r ∈ Z} is a subgroup of (Z, +), the
integers under addition.
(SG1) We have 1 = ζ 0 ∈ Cn .
(SG2) By the rules for products of powers, ζ r ζ s = ζ r+s ∈ Cn .
(SG3) Clearly ζ −r is the inverse of ζ r .
(d) The special linear group SLn (R) (matrices of determinant 1) is a sub-
group of GLn (R) (invertible matrices, with operation matrix multi-
plication).
72
(SG1) The identity matrix In has determinant 1.
(SG2) We know that det(AB) = det(A) det(B), so if A, B have
determinant 1, so does AB.
(SG3) We also know that det(A−1 ) = det(A)−1 , so if A has determ-
inant 1, so does A−1 .
For n = 1, GL1 (R) = R∗ = R\{0}; SL1 (R) = {1} is trivial and proper.
For any n > 1, SLn (R) is a proper non-trivial subgroup of GLn (R).
Sometimes it is instructive to see non-examples too, as in this case, to
see why all three conditions in the subgroup test are necessary.
Examples B.26.
(a) The set of odd integers, 2Z + 1 = {2r + 1 | r ∈ Z}, is not a subgroup
of (Z, +).
73
Exercise B.28. Let (G, ∗) be a group and consider the set
m1 Z ∩ m2 Z ∩ · · · ∩ mr Z = lcm(m1 , m2 , . . . , mr )Z.
The collection H T
= {mZ | m ∈ N} is a N-indexed family of subgroups of
Z. The intersection m∈N mZ consists of all
T integers that are divisible by
every m ∈ N. Only 0 has this property, so m∈N mZ = {0} is trivial.
Sometimes we have a subset S of a group G that is not a subgroup, but
where we would like to find a subgroup of G that contains S. Obviously we
could take all of G but this won’t tell us much: we really want the smallest
subgroup of G containing S.
“Smallest” here might mean “smallest size”, but for infinite sets it is
better to think of smallest as “smallest under inclusions of sets”. But we
know how to find the smallest set: it is the intersection. (The intersection is
a subset of every set in the collection and is the unique smallest such under
inclusion.) This leads us to make the following definition.
Definition B.33. Let G be a group and let S ⊆ G be a subset of G. Let
S = {H ≤ GT| S ⊆ H} be the family of subgroups of G that contain S. The
intersection S is a subgroup ofTG containing S, called the subgroup of G
generated by S. We write hSi = S.
74
Notice that by Proposition B.29 (or rather the slightly more general
version in the exercise that follows) hSi is a subgroup of G. It is also the
unique subgroup having the desired property, namely that it contains S. If
S is actually a subgroup already, then hSi = S.
Example B.34. In (Z, +), we have hmi = mZ (and in particular h1i = Z)
and h{m, n}i = hcf(m, n)Z, where hcf denotes the highest common factor.
This key example leads us to the following definition.
That is, cyclic groups are 1-generator, i.e. can be generated by a single
element. With a little work, one can show that a finite cyclic group is
necessarily isomorphic to Zn for some n and any infinite cyclic group is
isomorphic to Z.
We will often use a “generic” cyclic group Cn , given by
Cn = hg | g n = ei
More than once we have used the notion of a presentation of a group, for
example when we described the dihedral groups and most recently the cyclic
groups. We can now say a little more precisely what this means, although
we will not check all the details.
Giving a presentation for a group G, written hX | Ri, is the claim that G
is isomorphic to a quotient of the free group generated by X, F(X) = hXi.
The free group on a set is analogous to forming polynomials in a set
of variables: the elements of the free group are words in the set X, which
75
simply means finite expressions of the form x±1 ±1 ±1 with x ∈ X.
1 x2 · · · xr i
This becomes a group by the operation of concatenation (sticking two such
expressions together), where we understand that the expression of length 0
is the identity e and xi x−1 −1
i = e = xi xi .
The particular quotient we take to form hX | Ri is the one where we take
R to be a set of elements of F(X), form the subgroup hRi they generate
and then take the normal closure of this, i.e. we find the smallest normal
def
subgroup hRi of F(X) containing hRi. Then hX | Ri = F(X)/hRi.
Elements of X are called generators of hX | Ri and elements of R are
called relators. In practice, we often give elements of R as relations rather
than relators. Defining by way of an example, if we want a relation such as
b−1 ab = a−1 (imposing a relation meaning that in the group we want the
two sides to be equal) then formally we should make this into the relator
b−1 aba and put this element in R. Then in the quotient group b−1 aba will
become equal to the identity element of the quotient, so b−1 aba = e, i.e.
b−1 ab = a−1 . (We are being sloppy and using the same letters for the
elements a, b of F(X) and their images a + hRi, b + hRi in the quotient, but
this is common practice.) So rather than Cn = hg | g n = ei we should write
hg | g n i, but the former is easier to understand and we do that instead.
B.3 Cosets
Throughout this section, we will consider a group G and a subgroup H
of G. As previously, we will mostly write the group operation in G by
concatenation: if g, h ∈ G, gh will denote the element obtained from g and
h (in that order) under the binary operation of the group. The exception
will be in groups where the binary operation is addition (such as (Z, +)),
where we will write g + h.
76
If H = {e, h1 , h2 , . . . , hr } is finite, this is clear: gH = {g, gh1 , gh2 , . . . , ghr }
and by Proposition B.8(iii) (the cancellation property for groups) ghi = ghj
would imply hi = hj , so the elements g, gh1 , . . . , ghr are distinct. In partic-
ular, |gH| = |H|. It is natural to think of gH as the “translations by g” of
the elements of H.
In fact, the claims of the above paragraph still hold if H is infinite, if we
work a little more abstractly.
So any two cosets of H have the same size. Notice too that eH = H so
H is a coset of itself (the “translation” of H that does nothing).
However it is very much not true that if g1 6= g2 then g1 H 6= g2 H;
shortly we will see the correct condition for when two cosets are equal. This
condition will also tell us when a coset can be a subgroup (sneak preview:
only the “trivial” coset eH = H is).
First, let us look at a very important example.
Example B.40. We consider (Z, +) and its (cyclic) subgroup 4Z. Since we
are working in an additive group, we write the group operation in our coset
notation, so the cosets of 4Z are r + 4Z = {r + m | m ∈ 4Z} = {r + 4n | n ∈
Z}.
We see that the left cosets of 4Z are precisely the congruence classes of
integers modulo 4. The congruence class modulo 4 containing 3, for example,
is the set of integers {. . . , −9, −5, −1, 3, 7, 11, . . . }, which is exactly the set
of integers of the form 3 + 4n, namely {3 + 4n | n ∈ Z} = 3 + 4Z. So cosets
in particular describe congruence classes of integers.
Notice that every integer belongs to some left coset and that no integer
belongs to more than one left coset. This is because the congruence classes
modulo n for some fixed n partition Z, since congruence modulo n is an
equivalence relation. The following theorem is a vast generalisation of this
result, which corresponds to the specific group (Z, +).
g RH k ⇐⇒ g −1 k ∈ H
is an equivalence relation.
(ii) The equivalence classes for RH are precisely the left cosets gH, g ∈ G.
77
That is, for g1 , g2 ∈ G, g1 H ∩ g2 H is either empty or g1 H S
= g2 H (and the
latter happens if and only if g1−1 g2 ∈ H), and furthermore g∈G gH = G.
78
{0, 1, 2, 3} as these are the minimal positive representatives, and we will
usually make this choice and take TnZ as above.
Note however that {−2, −1, 0, 1, 2} is another natural choice for a trans-
versal of 5Z, and a similar choice can be made for (2m + 1)Z (i.e. odd n).
Since this works less well for even n, it is less commonly used, until one
wants to consider pZ for p prime, when in all but one case p is odd.
You will probably have noticed that we have been saying “left cosets”
and that the definition of gH is asymmetric. Unsurprisingly, one can make
the corresponding definition of a right coset.
79
gH = Hg for all g ∈ G, we can define a binary operation on cosets by
TH
g4H g4 g9 g 14 g 19 g 24 g 29
g3H g3 g8 g 13 g 18 g 23 g 28
g2H g2 g7 g 12 g 17 g 22 g 27
g1H g1 g6 g 11 g 16 g 21 g 26
g0H g0 g5 g 10 g 15 g 20 g 25
80
Corollary B.54. Let G be a finite group of order |G| = n. Then for all
g ∈ G, o(g) n and g n = e.
Corollary B.55. A group of prime order is cyclic and has no proper non-
trivial subgroups. Any non-identity element generates the group.
Groups with no proper non-trivial normal subgroups are called simple groups.
At the end of our work on groups, we will see in what sense these groups are
“simple” (they can be extremely large and complicated by other measures)
and why they are the building blocks for all groups. Cyclic groups of prime
order are very simple: they have no proper non-trivial subgroups at all!
Proposition B.56. If H, K ≤ G and hcf(|H|, |K|) = 1, then H ∩ K = {e}.
B.5 Homomorphisms
When we have two groups G and H and we want to relate the group struc-
ture on G to that on H, to compare properties between them, the right
mathematical thing to do is to start with a function ϕ : G → H. This gives
us a relationship between the underlying sets but this need not relate the
group structures: we need ϕ to be compatible with the binary operations
on G and H.
This leads us to the following definition.
Definition B.57. Let (G, ∗) and (H, ◦) be groups. A function ϕ : G → H
is called a group homomorphism if
for all g1 , g2 ∈ G.
Notice in particular that the operation in G, ∗, is being used on the left-hand
side of this equation (on g1 , g2 ∈ G), and that the operation in H, ◦ is being
used on the right-hand side (on ϕ(g1 ), ϕ(g2 ) ∈ H). This is an instance when
it is helpful to write the group operations explicitly.
First, let us deal with some elementary properties of homomorphisms.
Proposition B.58. Let ϕ : G → H be a group homomorphism. Then
(i) ϕ(eG ) = eH ;
(ii) ϕ(g −1 ) = ϕ(g)−1 for all g ∈ G;
(iii) ϕ(g n ) = ϕ(g)n for all g ∈ G, n ∈ Z and
(iv) o(ϕ(g)) | o(g) for all g ∈ G.
The composition of two functions is again a function and we would expect
that if two composable functions preserved group structure, their composi-
tion would too. Indeed this is the case.
81
Proposition B.59. Let ϕ : G → H and σ : H → K be group homomorph-
isms. Then σ ◦ ϕ : G → K is a group homomorphism.
Examples B.60.
for all x, y ∈ R∗ .
for all x, y ∈ R. (Note the two different notations for the group oper-
ations!)
(d) Recall from Remark B.19 that we have a function ϕ : Sn → GLn (R),
ϕ(σ) = Aσ where (
1 if σ(j) = i
(Aσ )ij =
0 otherwise
is the permutation matrix associated to σ.
Then Aστ = Aσ Aτ , where the left-hand side involves the composition
of permutations and the right-hand side is the usual multiplication of
matrices. (The definition of Aσ , with a 1 when σ(j) = i might seem
to be the “wrong way round”; however if we made the more natural-
looking definition with σ(i) = j, we would have an order reversal in
the products.) It follows that ϕ : Sn → GLn (R), ϕ(σ) = Aσ is a group
homomorphism. For an explicit example, look back to Remark B.19.
Exercise B.61. Let (G, ∗) and (H, ◦) be groups with identity elements eG and
eH respectively. Show that the function ϕ : G → H defined by ϕ(g) = eH
for all g ∈ G is a group homomorphism. We call this the trivial group
homomorphism.
82
Recall that certain sorts of functions are special, namely injective, sur-
jective and bijective functions; also, a function is bijective if and only if
it is invertible. If a group homomorphism has one of these properties, we
(sometimes) use a special name for it. So an injective group homomorphism
is also known as a monomorphism and a surjective group homomorphism is
called an epimorphism. This comes from the wider category theory context,
but in Grp being mono is equivalent to being injective and being epi is the
same as being surjective.
Examples B.62.
but 3 ∈
/ Im ϕ.
and Im ϕ = {z ∈ C | kzk = 1} ( C∗ .
83
Note that the identity function idG : G → G, idG (g) = g is a group
homomorphism13 and is a bijection, so is a group isomorphism. This is no
surprise: every group should have the same group structure as itself!
A group isomorphism is a particular kind of homomorphism—a function
with certain properties. We can use these to talk about how two groups
might be related, as follows.
Definition B.64. We say that two groups G and H are isomorphic if and
only if there exists a group isomorphism ϕ : G → H. If G and H are
isomorphic, we often write G ∼
= H, for short.
Lemma B.65. Let G and H be groups. Then the following are are equival-
ent:
84
Example B.66. The groups C6 and S3 are not isomorphic. They have the
same order (|C6 | = 6 = 3! = |S3 |) but C6 is Abelian and S3 is not. Indeed,
C6 is cyclic and S3 is not (if it were, every permutation in S3 would have to
be an r-cycle for some r, but S3 has both transpositions and 3-cycles).
However, sharing some properties is not (usually) enough to prove that
the groups are isomorphic: almost always you will need to find an explicit
isomorphism between them. That said, if two groups share lots of properties,
this might reasonably lead you to conjecture that they are isomorphic; this
would not be a proof, though.
Example B.67. Recall that the subset C12 = {e2kπi/12 | 0 ≤ k < 12} of C∗ is
a cyclic subgroup, so is a cyclic group in its own right. We may write it as
C12 = hζi for ζ = e2πi/12 .
Let G = hgi be the (generic) cyclic group of order 12. So G = {e, g, g 2 , . . . , g 11 }
with group operation g r g s = g r+s . We claim that C12 is isomorphic to G.
Let ϕ : C12 → G be defined by ϕ(ζ k ) = g k for all 0 ≤ k < 12. Clearly ϕ
is a bijection and
so ϕ is a homomorphism.
Now consider (Z12 , +), the integers modulo 12, with addition (modulo
12). So Z12 = {b0, b
1, . . . , 11}
b with b c if and only if a + b ≡ c mod 12.
a + bb = b
We claim that this group is also isomorphic to G, and hence is isomorphic
to C12 too.
The function we need is ψ : Z12 → G, ψ(b a) = g a . Then ψ is a bijection:
a b
g = g if and only if a ≡ b mod 12. It is a group homomorphism since
So ψ is an isomorphism.
The observant among you will have noticed that the number 12 is irrel-
evant in this example: everything works for a general n ∈ N. We can extend
it further, as follows.
Note that neither the statement or the proof assume that the cyclic
groups are finite.
As we said before, two groups having the same order is far from being
enough to tell us that the groups are isomorphic. The message we take from
this proposition is that being cyclic is a very strong condition: same order
plus cyclic is enough to give us isomorphic.
85
Previously we noted that the identity map provides an isomorphism of
a group with itself. In fact, a group can have other self-isomorphisms and
it is often an important question to know how many. In some sense, these
are the symmetries of the group (which itself may be the symmetries of
something!).
Definition B.69. Let G be a group. A group isomorphism ϕ : G → G is
called an automorphism of G.
Lemma B.70. Let G be a group. The set of automorphisms of G forms a
group under composition.
This justifies the following definition.
Definition B.71. Let G be a group. The group of automorphisms of G,
AutGrp (G) = {ϕ : G → G | ϕ is an isomorphism} is called the automorph-
ism group of G.
Example B.72. One can show that AutGrp (Cn ) ∼ = (Z×n , ×), the group of units
of the ring of integers modulo n. In particular, we have AutGrp (C2 ) = {e}
and AutGrp (C3 ) ∼= C2 .
We will not justify these claims here: finding the automorphism group
of even a small group usually requires more sophisticated technology than
we currently have to hand. We will simply note that the above has an
important relationship with number theory: n 7→ |AutGrp (Cn )| is called the
Euler totient function14 and is a very important function about which much
more is said in MATH328 Number Theory.
Instead, we will move on and look at some important subsets of a group
associated to homomorphisms.
86
Definition B.73. Let ϕ : G → H be a group homomorphism. The kernel
of ϕ is defined to be the subset of G given by
Ker ϕ = {g ∈ G | ϕ(g) = eH }.
Im ϕ = {h ∈ H | (∃ g ∈ G)(ϕ(g) = h)}.
Notice that the definition of the kernel would not be possible without the
identity element eH and that the image is exactly the image of ϕ as a
function. Be very clear in your mind that Ker ϕ is a subset of the domain,
G, and Im ϕ is a subset of the codomain, H.
You might find the following picture helpful:
Ker ϕ Im ϕ
eG eH
ϕ
G H
Very shortly we will come to examples but first we will prove some basic
facts about kernels and images that will be very helpful in working out the
examples.
87
(iii) The homomorphism ϕ is injective if and only if Ker ϕ = {eG }.
(iv) The homomorphism ϕ is surjective if and only if Im ϕ = H.
(v) The homomorphism ϕ is an isomorphism if and only if Ker ϕ = {eG }
and Im ϕ = H.
Remark B.76. Let ϕ : G → H be an injective homomorphism. Then we can
consider the function ψ = ϕ|Im ϕ : G → Im ϕ, the codomain restriction to
Im ϕ of ϕ. (This is the function that takes the same values as ϕ but where
we “throw away” any elements of H not in the image of ϕ, and so just take
Im ϕ as the codomain.) Since Ker ψ = Ker ϕ = {eG } and Im ψ = Im ϕ, ψ is
both injective and surjective. Hence ψ is an isomorphism of G with Im ϕ.
Since Im ϕ is a subgroup of H, in this situation we say that G is iso-
morphic to a subgroup of H.
That is, to show that a group G is isomorphic to a subgroup of a group
H, we show that there exists an injective homomorphism from G to H. If
that homomorphism is also surjective, we have that G is isomorphic to H
itself.
Let us revisit the examples from Examples B.60 and B.62.
Examples B.77.
(a) The homomorphism ϕ : Z → Z, ϕ(n) = 2n for all n ∈ Z is injective so
Ker ϕ = {eZ } = {0}.
The image of ϕ is all integers of the form 2n, so Im ϕ = 2Z.
(b) The homomorphism ϕ : R∗ → R∗ , ϕ(x) = x2 for all x ∈ R∗ is not
injective or surjective. Its kernel is all non-zero real numbers whose
square is equal to 1 (which is the identity element in R∗ = (R\{0}, ×)).
So Ker ϕ = {1, −1}.
The image of ϕ is the set of all non-zero real numbers that are squares;
there is no “nice” description of this, so we simply have
Im ϕ = {y ∈ R∗ | y = x2 for some x ∈ R∗ }.
Im ϕ = {z ∈ C | kzk = 1} ( C∗ .
88
(e) The homomorphism Σ : Sn → {1, −1}, Σ(π) = sign(π) for all π ∈ Sn
is not injective. Its kernel is
Ker Σ = {σ ∈ Sn | sign(σ) = 1}
since 1 is the identity element in ({1, −1}, ×). By definition, these are
the even permutations, An , so Ker Σ = An . Since there exist both
even and odd permutations, Σ is surjective and Im Σ = {1, −1}.
gK = {l ∈ G | g Rϕ l} = {l ∈ G | ϕ(g) = ϕ(l)}.
This is the analogous result for groups to the Dimension Theorem for vector
spaces, which asserts that for a linear transformation T : V → W we have
dim V = dim Ker T + dim Im T .
In fact, both of these are “numerical shadows” of stronger statements:
shortly we will introduce quotient groups and prove the stronger version for
groups.
Since Ker ϕ ≤ G and Im ϕ ≤ H, we already knew that |Ker ϕ| divides |G|
and |Im ϕ| divides |H|, by Lagrange’s theorem. However this corollary tells
us that |Im ϕ| also divides |G|, which can be useful to know, in applications
such as the following.
Example B.80. Let G be a group of order 16 and H a group of order 9.
Then the only homomorphism ϕ : G → H is the trivial homomorphism with
ϕ(g) = eH for all g ∈ G. For if ϕ : G → H is a homomorphism, |Im ϕ| must
divide |G| and |H|, but hcf(16, 9) = 1 so |Im ϕ| = 1 and so Im ϕ = {eH }.
89
B.7 Cayley’s theorem
Sometimes one needs to calculate in particular explicit examples. This might
be done by hand or, more commonly now, by a computer package. But in
order to do so, we need a “nice” way to represent the elements of our group,
to store them and to calculate the result of the binary operation on them. In
this section, we will see two theoretical results that can be used to provide
such representations. The first is called Cayley’s theorem.
Definition B.81. Let X be a set. The symmetric group on X is defined to
be the group SX = (Bij(X), ◦) of bijections between X and itself.
Theorem B.82 (Cayley’s theorem). Let G be a group. There is an injective
homomorphism λ : G → SG .
Corollary B.83. Every group G is isomorphic to a subgroup of SG .
In particular, every finite group G of order n is isomorphic to a subgroup
of the symmetric group of degree n, Sn .
Corollary B.84. Let n ∈ N. There are finitely many non-isomorphic
groups of order n.
The phrase “up to isomorphism” at the end here is very important: there
are infinitely many groups with one element, for example, but they are all
isomorphic15 .
We can also join two previous results together: Cayley’s theorem and the
homomorphism from the symmetric group to the general linear group given
by taking permutation matrices. This gives us, for free, a matrix version of
Cayley’s theorem.
Theorem B.85. Let G be a finite group. There is an injective homomorph-
ism µ : G → GL|G| (R).
90
The ingredients for the construction are:
• a group (G, ∗) (to try to avoid confusion, we will revert to using an
explicit symbol to denote the binary operation in G);
91
Corollary B.94. Let G be a group and N a normal subgroup of G. Then N
is the kernel of a group homomorphism with domain G, namely the quotient
homomorphism π : G → G/N .
We can illustrate the maps in this theorem via the following diagram,
known as a “commutative diagram” (the “commutative” refers to the fact
that following along the arrows in either possible way gives the same result).
ϕ
G H
π
∃!ϕ̄
G/N
A consequence of this theorem is that there is a one-to-one correspond-
ence between homomorphisms ϕ : G → H with N ⊆ Ker ϕ and homomorph-
isms ϕ̄ : G/N → H, the correspondence being given by composing with π.
Since quotient groups can be complicated to understand, this tells us that
to find homomorphisms whose domain is G/N , we can instead look for ho-
momorphisms whose domain is G and just check if N is a subgroup of the
kernel of these. (Depending on the situation, the converse direction can be
very helpful too, of course.)
The above theorem has done much of the heavy lifting to enable us to
state and prove in a nice, clean fashion the First Isomorphism Theorem.
(Arguably, it should be called the “First Isomorphism Corollary” since it
follows essentially immediately from the previous theorem, but often the-
orems are called theorems for their significance rather than their outright
difficulty.)
G/ Ker ϕ ∼
= Im ϕ
a × (b + c) = (a × b) + (a × c) and (a + b) × c = (a × c) + (b × c)
92
for all a, b, c ∈ R. The identity element for the group (R, +) will be denoted
0R . The inverse of a with respect to + will be denoted by −a.
Remark B.98. The distributive laws can seem a little like they appear out
of thin air. However, if one stares at them for a while, one sees that in fact
they say that left and right multiplication are group homomorphisms. More
precisely, given a ∈ R, the function µL L
a : R → R given by µa (b) = a × b is a
homomorphism from the Abelian group R = (R, +) to itself, and similarly
for µR R
c : R → R, µc (b) = b×c. From this point of view, the conditions above
are not so unnatural.
Definition B.99. Let R be a ring. The set S(R) of (N0 -indexed) sequences
of elements of R, with the operations +S(R) and ×S(R) defined by
(a +S(R) b)n = an +R bn
and X
(a ×S(R) b)n = ai ×R bj
i+j=n
is a ring, called the ring of formal power series (in one variable) over R. We
denote this ring by R[[x]] and the ring R is called the base ring of R[[x]].
Remark B.100. If R has a multiplicative identity 1R (see Definition B.107)
then the “variable” x may be identified with the sequence
(
1R for n = 1
xn =
0R otherwise.
This covers most familiar situations but has the perhaps surprising implic-
ation that if R does not have a multiplicative identity, then x = 0R + 1R x +
0R x2 + 0R x3 + · · · ∈
/ R[[x]].
Now that we have the ring of formal power series, it is straightforward to
define the ring of polynomials. Polynomials are just “special” power series,
namely they are the power series that are “eventually zero”, i.e. after some
point, every element of the sequence is 0R . Another way to say this is that
only finitely many elements of the corresponding sequence are non-zero. We
say that a sequence a is finitely supported if it has this property: a is finitely
supported if |{i | ai 6= 0R }| is finite.
Definition B.101. Let R be a ring. The set Sfs (R) of finitely supported
(N0 -indexed) sequences of elements of R, with the operations =Sfs (R) and
×Sfs (R) defined by
(a +Sfs (R) b)n = an +R bn
and X
(a ×Sfs (R) b)n = ai ×R bj
i+j=n
93
is a ring, called the ring of polynomials (in one variable) over R. We denote
this ring by R[x] and the ring R is called the base ring of R[x].
Remark B.102. As in Remark B.100, if R has a multiplicative identity 1R
then the “variable” x may be identified with the (finitely supported) se-
quence (
1R for n = 1
xn =
0R otherwise.
Again this leads to the counter-intuitive observation that if R does not have
a multiplicative identity, then x ∈/ R[x], i.e. x is not a polynomial—the
point being that it is not a polynomial with coefficients in R. For a concrete
example, consider the ring 2Z[x] of polynomials in one variable x with even
integer coefficients: x ∈
/ 2Z[x].
Definition B.103. Let R be a ring and R[x] the polynomial ring in one
variable over R. Let p = (pn ) be a non-zero element of R[x], i.e. there exists
i such that pi 6= 0R .
The degree of p is defined to be
deg p = max{k | pk 6= 0R }.
94
(ii) For all a ∈ R, 0R × a = 0R = a × 0R .
(b) The ring (Z, +, ×), with the usual addition and multiplication of in-
tegers, has multiplicative identity 1; similarly for Q, R and C, with
their usual operations.
(d) The matrix rings Mn (Z), Mn (Q), Mn (R) and Mn (C) all have multi-
plicative identities, namely the n × n identity matrix In ,
(
1 if i = j
(In )ij =
0 otherwise.
The matrix ring Mn (R), for R an arbitrary ring, need not have a
multiplicative identity; shortly we will see precisely when it does.
(e) The ring of functions F(Y, R) (respectively F(Y, C)) (where Y is any
set) has a multiplicative identity, the constant function 1 : Y → R
(respectively 1 : Y → C) defined by 1(y) = 1 for all y ∈ Y .
(f) The polynomial rings Z[x], Q[x], R[x] and C[x] have multiplicative
identities, namely the constant polynomial 1 = 1 + 0x + 0x2 + · · · .
The polynomial ring R[x], for R an arbitrary ring, need not have a
multiplicative identity; again we shall shortly see precisely when it
does.
95
Just as in the case of groups, rings with a commutative multiplication
operation are particularly special. Remember that a ring (R, +, ×) is by
definition an Abelian group with respect to +, so a + b = b + a for all a, b
always. The issue at hand for rings is whether or not the multiplication
operation × is commutative.
Definition B.109. Let (R, +, ×) be a ring. We say that R is a commutative
ring if the multiplication operation × is commutative. That is, a × b = b × a
for all a, b ∈ R.
If R is not commutative, we say R is non-commutative.
Note that the negation of (∀a, b ∈ R)(a × b = b × a) is
¬ ((∀a, b ∈ R)(a × b = b × a)) = (∃ a, b ∈ R)(a × b 6= b × a),
so that to show a ring R is non-commutative we must find two distinct
(specific) elements of R whose products a × b and b × a are not equal. (We
must have a 6= b, since a × a = a × a, of course.)
We have already seen a number of examples of commutative and non-
commutative rings.
Examples B.110.
(a) As you have known for a long time, Z, Q, R and C are commutative
rings.
(b) The integers modulo n, Zn , are a commutative ring.
(c) Matrix rings are usually non-commutative.
(d) The rings of functions F(Y, R) and F(Y, C) are commutative: multi-
plication is defined pointwise so if f, g ∈ F(Y, R),
(f g)(y) = f (y)g(y) = g(y)f (y) = (gf )(y)
for all y ∈ Y ; similarly for f, g ∈ F(Y, C). A little careful thought
shows that the key issue here for F(Y, R) with R an arbitrary ring is
whether or not R is commutative.
(e) Polynomial rings R[x] can be commutative or not, depending on R.
(f) The ring of even integers 2Z is commutative, because its multiplica-
tion is that of Z, restricted to the even integers, and multiplication of
integers is commutative.
Lemma B.111. Let R be a ring such that there exist r, s ∈ R with rs 6= 0R .
Then the matrix ring Mn (R) is commutative if and only if n = 1 and R is
commutative. That is, Mn (R) is non-commutative if n ≥ 2, and M1 (R) is
non-commutative if and only if R is.
Lemma B.112. Let R be a ring. Then the polynomial ring R[x] is com-
mutative if and only if R is.
96
B.12 Subrings
Just as with groups and subgroups, it is natural to study subsets of rings
that are themselves rings. Indeed, in a few places above it would have been
natural to say “subring”, so it is not before time that we give the definition.
Definition B.113. Let (R, +, ×) be a ring. A subset S of R is said to be
a subring if (S, +, ×) is a ring. If this is the case, we write S 6 R.
Proposition B.114 (Subring test). Let S be a subset of a ring (R, +, ×)
and denote the additive identity element of R by 0R . Then S is a subring of
R if and only if the following two conditions are satisfied:
(SR1) S is a subgroup of (R, +);
(SR2) for all s1 , s2 ∈ S, we have s1 × s2 ∈ S.
Lemma B.115. Every subring of a commutative ring is commutative.
Exercise B.116. Let R and S be rings, and let T denote their Cartesian
product, that is,
T = R × S = {(a, b) | a ∈ R, b ∈ S}.
(a, b) +T (c, d) = (a +R c, b +S d)
and
(a, b) ×T (c, d) = (a ×R c, b ×S d)
for a, c ∈ R and b, d ∈ S.
(a) Show that T is a ring with respect to these operations.
(b) Show that T is commutative if and only if R and S are commutative.
B.13 Homomorphisms
Earlier, in Section B.5, we saw that the correct way to relate two groups
(G, ∗) and (H, ◦) is to define a group homomorphism, this being a function
ϕ : G → H such that ϕ(g1 ∗ g2 ) = ϕ(g1 ) ◦ ϕ(g2 ) for all g1 , g2 ∈ G. A
ring (R, +, ×) is by definition an Abelian group (R, +) with a compatible
multiplication ×, so a ring homomorphism must be a group homomorphism
that also preserves the multiplication operations.
Definition B.117. Let (R, +R , ×R ) and (S, +S , ×S ) be rings. A function
ϕ : R → S is called a ring homomorphism if ϕ is a group homomorphism
and preserves the multiplication operations. Explicitly,
97
(H1) ϕ(r1 +R r2 ) = ϕ(r1 ) +S ϕ(r2 ) and
for all r1 , r2 ∈ R.
(i) ϕ(0R ) = 0S ;
Example B.120. Next we give the ring analogue of the trivial group homo-
morphism defined in Exercise B.61.
Let R and S be rings and let ϕ : R → S be the function defined by
ϕ(a) = 0S for all a ∈ R. Then ϕ is a ring homomorphism because for all
a, b ∈ R, we have
98
Again, the identity function idR : R → R, idR (r) = r is a ring homo-
morphism16 and is a bijection, so is a ring isomorphism.
Definition B.123. We say that two rings R and S are isomorphic if and
only if there exists a ring isomorphism ϕ : R → S. If R and S are isomorphic,
we often write R ∼ = S.
Lemma B.124. Let R and S be rings. Then the following are are equival-
ent:
If R and S are unital rings, we may or may not have that a ring homo-
morphism sends 1R to 1S . If we want to be sure of this, we need to ask for
it explicitly:
Definition B.127. Let R and S be unital rings. We say that a ring homo-
morphism ϕ : R → S is unital if ϕ(1R ) = 1S .
Ker ϕ = {r ∈ R | ϕ(r) = 0S }.
16
The collection of all rings together with all ring homomorphisms between them also
forms a category.
99
Definition B.129. Let ϕ : R → S be a ring homomorphism. The image of
ϕ is defined to be the subset of S given by
+I : LI × LI → LI , (r + I) +I (s + I) = (r + s) + I
and
×I : LI × LI → LI , (r + I) ×I (s + I) = (r × s) + I
define binary operations on LI .
100
(i) The operation +I defines a group structure on LI , the set of left cosets
of I in R (with respect to addition).
101
Theorem B.144 (First Isomorphism Theorem for Rings). Let R and S be
rings and let ϕ : R → S be a ring homomorphism. Then
R/ Ker ϕ ∼
= Im ϕ.
where [T ]B
B = (tij ).
Eigenvectors and eigenvalues will be important for us, too. An eigen-
vector for a linear map T : V → V is a vector v ∈ V such that there exists
λ ∈ K such that T v = λv. The scalar λ is called the eigenvalue associated
to the eigenvector v.
To find eigenvalues, we may use the following result.
Theorem B.145. Let T : V → V and set M = [T ]B B for some choice of
basis B for V . Then for any λ ∈ K there exists an eigenvector v for T with
eigenvalue λ if and only if λ is a root of the characteristic polynomial of T ,
det(M − tI).
Then to find eigenvectors, we solve the simultaneous linear equations
M v = λv. (Fast computational methods to compute eigenvectors and ei-
genvalues exist but is also important to know how to do this by hand.)
102
The subset of eigenvectors for a fixed eigenvalue λ forms a subspace of
V , which we denote Vλ and call the λ-eigenspace.
If a vector space V has a basis consisting of eigenvectors for a linear
transformation T : V → V , then we say T is diagonalisable, for then with
respect to this basis, T is represented by a diagonal matrix with the eigen-
values on the diagonal. It follows that if T is diagonalisable, its trace is equal
to the sum of its eigenvalues and its determinant is equal to the product of
its eigenvalues.
References
[ASS] Ibrahim Assem, Daniel Simson, and Andrzej Skowroński. Elements
of the representation theory of associative algebras. Vol. 1, volume 65
of London Mathematical Society Student Texts. Cambridge Univer-
sity Press, Cambridge, 2006. 978-0-521-58423-4; 978-0-521-58631-3;
0-521-58631-3. Techniques of representation theory.
103
[Sch] Ralf Schiffler. Quiver representations. CMS Books Math./Ouvrages
Math. SMC. Cham: Springer, 2014. 978-3-319-09203-4; 978-3-319-
09204-1.
104