Lecture_notes_algebra_geometry
Lecture_notes_algebra_geometry
Andrea Ferraguti
September 6, 2023
Disclaimer: If you find any mistake and/or typo in these lecture notes you
can let me know at [email protected].
ii
Contents
Chapter 0: Basics 1
0.1 Sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
0.2 Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
0.3 The induction principle . . . . . . . . . . . . . . . . . . . . . . 7
0.4 Real and complex numbers . . . . . . . . . . . . . . . . . . . . 9
0.5 Polynomials . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
0.6 Matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
iii
6.6 Relative position of subspaces via equations . . . . . . . . . . 142
6.7 Pencils and bundles of lines and planes . . . . . . . . . . . . . 150
iv
v
Andrea Ferraguti Chapter 0: Basics
Chapter 0: Basics
0.1 Sets
Throughout these lecture notes, a set will be an unordered collection of ob-
jects, without repetitions. The objects contained in a set are called elements.
Elements of a set will be enclosed between curly brackets.
1
Andrea Ferraguti Chapter 0: Basics
• ∀x ∈ Z, ∃! y ∈ Z such that x + y = 0.
• @x ∈ N such that x2 = 2.
• ∀x ∈ Z, x2 ≤ 4 ⇐⇒ −2 ≤ x ≤ 2
Remark 0.1.4. Every set S has at least two subsets: the empty set and S
itself.
For S, T sets, the following operations are allowed and produce a new set.
• The intersection of S, T is the set:
S ∩ T := {s : s ∈ S and s ∈ T }.
S ∪ T := {s : s ∈ S or s ∈ T }.
S \ T := {s : s ∈ S and s ∈
/ T }.
2
Andrea Ferraguti Chapter 0: Basics
1. S ∩ S = S ∪ S = S;
2. S ∩ ∅ = ∅;
3. S ∪ ∅ = S;
4. (S ∩ T ) ∩ U = S ∩ (T ∩ U );
5. (S ∪ T ) ∪ U = S ∪ (T ∪ U );
6. S ⊆ T ⇐⇒ S ∩ T = S ⇐⇒ S ∪ T = T ;
7. S ∩ (T ∪ U ) = (S ∩ T ) ∪ (S ∩ U );
8. S ∪ (T ∩ U ) = (S ∪ T ) ∩ (S ∪ U ).
S × T := {(s, t) : s ∈ S, t ∈ T }.
S × T = {(a, 1), (a, 2), (a, 3), (b, 1), (b, 2), (b, 3)}.
3
Andrea Ferraguti Chapter 0: Basics
If the n sets S1 , S2 , . . . , Sn are all equal to a set S, we write S n for the cartesian
product S1 × . . . × Sn .
0.2 Functions
Let S, T be sets.
√
Example 0.2.2. If S = {0, 1} and T = {a, ♠, 7}, the following are
correspondences between S and T :
1. S × T ;
2. ∅;
f: Z→N
4
Andrea Ferraguti Chapter 0: Basics
x 7→ x2
means that f is the function between Z and N that associates to every
integer x then natural number x2 .
Example 0.2.4.
5
Andrea Ferraguti Chapter 0: Basics
Note:-
What Proposition 0.2.6 says is that in order to compute the cardinality of
a union of finite sets, you must sum the cardinality of each single set, then
subtract the cardinality of all intersections of two of them, then add the
cardinality of all intersections of three of them, and so on. For example, if
n = 2 then:
|S1 ∪ S2 | = |S1 | + |S2 | − |S1 ∩ S2 |,
and if n = 3 then
|S1 ∪ S2 ∪ S3 | =
= |S1 | + |S2 | + |S3 | − |S1 ∩ S2 | − |S1 ∩ S3 | − |S2 ∩ S3 | + |S1 ∩ S2 ∩ S3 |.
If S1 , . . . , Sn are pairwise disjoint, then |S1 ∪. . .∪Sn | = |S1 |+|S2 |+. . .+|Sn |.
6
Andrea Ferraguti Chapter 0: Basics
then it follows that P (n) is true for every natural number n. We will now
illustrate this principle with three very classical examples.
Pn n(n+1)
Proposition 0.3.1. Let n be a natural number. Then i=0 i= 2
.
7
Andrea Ferraguti Chapter 0: Basics
We therefore need to prove two facts. The first one is that P (0) holds
true, i.e. we need to show that
0
X 0(0 + 1)
i= .
i=0
2
Proposition 0.3.3. Let S be a finite set with |S| = n. Then S has exactly
2n subsets.
8
Andrea Ferraguti Chapter 0: Basics
9
Andrea Ferraguti Chapter 0: Basics
One can verify that addition and multiplication are commutative and associa-
tive.
There is a bijective map, called conjugation, defined as follows:
·: C → C
a + bi 7→ a + bi := a − bi
Conjugation has the property that x + y = x + y and xy = x · y for every
x, y ∈ C. Moreover,
R = {x ∈ C : x = x}. (1)
0.5 Polynomials
Let K be a field (for the definition of field, see Definition 1.1.13). If you are
not familiar yet with the concept of field, just think of K as to one among
Q, R or C. A polynomial with coefficients in K is an expression of the form
p(x) = a0 + a1 x + a2 x2 + . . . + an xn ,
10
Andrea Ferraguti Chapter 0: Basics
and !
2n
X X
p(x)q(x) = ai b j xk .
k=0 i,j : i+j=k
When we add up two polynomials, the degree of the sum is at most the largest
degree between the two, namely
and equality holds if deg p(x) 6= deg q(x). When we multiply two polyomials,
their degrees add up:
p(z) := a0 + a1 z + a2 z 2 + . . . + an z n ,
11
Andrea Ferraguti Chapter 0: Basics
0.6 Matrices
Let K be a field (for the definition of field, see Definition 1.1.13). If you are
not familiar yet with the concept of field, just think of K as to one among
Q, R or C. Let m, n ≥ 1 be integers.
A : {1, . . . , m} × {1, . . . , n} → K.
For practical purposes, we identify a matrix with its image, and we arrange
the values of the function defining the matrix in a rectangular table with m
rows and n columns. That is, an m × n matrix A is a table of the following
12
Andrea Ferraguti Chapter 0: Basics
form:
a11 a12 . . . a1n
a21 a22 . . . a2n
A=
.
... ... ... ...
am1 am2 . . . amn
To ease the notation, we sometimes will write A = (aij ) i=1,...,m to say that A
j=1,...,n
is an m × n matrix whose entry in row i and column j is aij .
The set of all m×n matrices with coefficients in K is denoted by Mm×n (K).
If m = n, we shorten the notation by just writing Mn (K). Matrices belonging
to Mn (K) are called square matrices of size n.
Any two elements of Mm×n (K) can be added, as follows. Given
a a12 . . . a1n b b12 . . . b1n
11 11
a21 a22 . . . a2n b21 b22 . . . b2n
A=
and B =
... ... ... ... ... ... ... ...
am1 am2 . . . amn bm1 bm2 . . . bmn
we let
a11 + b11 a12 + b12 ... a1n + b1n
a21 + b21 a22 + b22 . . . a2n + b2n
A + B :=
.
... ... ... ...
am1 + bm1 am2 + bm2 . . . amn + bmn
Matrices could, in principle, be multiplied entry-by-entry, similarly to the way
we add them. However, this multiplication will not be used anywhere in these
lecture notes. On the other hand, we will now define a matrix multiplication
that can be performed only between an m × n and an n × p matrix, where
m, n and p are positive integers. If A = (aij ) i=1,...,m ∈ Mm×n (K) and B =
j=1,...,n
(bij ) i=1,...,n ∈ Mn×p (K), the matrix AB is the matrix (cij ) i=1,...,m ∈ Mm×p (K)
j=1,...,p j=1,...,p
where for every i ∈ {1, . . . , m} and every j ∈ {1, . . . , p}:
n
X
cij = aik bkj .
k=1
In other words, to find the entry in row i and column j of the matrix AB, we
need to take the i-th row of A (that has n entries) and the n-th column of B
13
Andrea Ferraguti Chapter 0: Basics
(that also has n entries), multiply the corresponding entries (the entry in the
i-th row and k-th column of A must be multiplied with the entry in the k-th
row and j-th column of B), and add up all the results.
Note:-
Given matrices A and B, the product AB makes sense only when the num-
ber of columns of A equals the number of rows of B.
Example 0.6.2.
1
2 3 1
• Let A := ∈ M2×3 (R) and B :=
2 ∈ M3×1 (R).
−1 0 1
2
10
Then AB = ∈ M2×1 (R).
1
2 2 1 1 −1
• Let A := ∈ M2 (Q) and B := ∈ M2×3 (Q).
0 0 0 0 3
2 2 4
Then AB = ∈ M2×3 (Q).
0 0 0
1
2 2
• The matrices A := ∈ M2 (C) and B := 2 ∈ M3×1 (C)
0 0
2
cannot be multiplied, since A has two columns and B has 3 rows.
14
Andrea Ferraguti Chapter 0: Basics
Example 0.6.5.
• The matrix
2 0 0 0
0 3 0
0
0 −1
0 0
√
0 0 0 − 2
is diagonal.
• The matrix
1 2 3
A 0 4 5
0 0 −1
is upper triangular.
15
Andrea Ferraguti Chapter 0: Basics
• The matrix
2 0 0 0
−7 3 0 0
0 1 −1 0
π 0 22 0
is lower triangular.
Example 0.6.8.
• Let
1 2 −3
A := .
0 1 7
Then
1 0
t
A = 2 1 .
−3 7
• Let
A := 1 2 3 4 .
16
Andrea Ferraguti Chapter 0: Basics
Then
1
t
2
A=
.
3
4
• Let
1 2 0
A := 5 6 7 .
√
π 8 6
Then
1 5 π
t
A = 2 6 8 .
√
0 7 6
• The matrix
1 0 −2
2 −3
0
−2 −3 0
is symmetric.
Remark 0.6.9.
17
Andrea Ferraguti Chapter 1: Vector spaces
Example 1.1.2.
18
Andrea Ferraguti Chapter 1: Vector spaces
1. ? is associative;
Example 1.1.10.
• The pair (Q, +) is an abelian group. The set (Q, ·) is not an abelian
group, because 0 does not possess an inverse.
19
Andrea Ferraguti Chapter 1: Vector spaces
{f : S → S s.t. f is bijective}
20
Andrea Ferraguti Chapter 1: Vector spaces
Example 1.1.14. Q, R, C are all fields with the usual operations of sum
and multiplication. Z is not a field with respect to the usual sum and
multiplication because elements different from ±1 do not have a multi-
plicative inverse.
The set F2 := {0, 1} is a field when endowed with the following two
operations:
+ 0 1 · 0 1
0 0 1 0 0 0
1 1 0 1 0 1
∗: K × V → V
21
Andrea Ferraguti Chapter 1: Vector spaces
4. for every v ∈ V , 1 ∗ v = v.
Note:-
Addition in K and addition in V are both denoted by +. However, beware of
the fact that these are different operations, since one is a function K × K →
K and the other one is a function V × V → V . Therefore, an expression of
the form α + v, where α ∈ K and v ∈ V , does not make any sense.
Note:-
While 0 denotes the zero vector, namely the neutral element of the group
(V, +), we denote by 0 the neutral element of the group (K, +). Hence these
are two very different objects, do not confuse them!
Example 1.1.17. Let V = {0} be the abelian group that possesses only
one element, the neutral element with respect to the operation + on V .
This is a K-vector space over any field K, the operation ∗ being defined
by α ∗ 0 = 0 for every α ∈ K. This is the simplest possible vector space,
although not a very interesting one.
There are several important examples of vector spaces, but for the sake of
these lecture notes the most important is by far the following one.
∗: K × V → V
22
Andrea Ferraguti Chapter 1: Vector spaces
v + w = (v1 + w1 , v2 + w2 , . . . , vn + wn )
and hence
Hence
as desired. It is a very useful exercise for the reader to verify that prop-
erties 2., 3. and 4. are satisfied as well.
Notice that when n = 1 we have V = K. When this happens, opera-
tion ∗ coincides with multiplication in K. Therefore, every field K, seen
as an abelian group with respect to the sum, is a vector space over K.
∗ : K × K[x]≤n → K[x]≤n
n
! n
X X
i
α, ai x 7→ (αai )xi
i=0 i=0
23
Andrea Ferraguti Chapter 1: Vector spaces
∗ : K × Mn (K) → Mn (K)
0 ∗ v = (0 + 0) ∗ v = 0 ∗ v + 0 ∗ v,
α ∗ 0 = α ∗ (0 + 0) = α ∗ 0 + α ∗ 0,
α ∗ 0 + 0 = α ∗ 0 + α ∗ 0,
24
Andrea Ferraguti Chapter 1: Vector spaces
that is, v = 0.
0 = 0 ∗ v = (1 − 1) ∗ v = 1 ∗ v + (−1) ∗ v = v + (−1) ∗ v,
25
Andrea Ferraguti Chapter 1: Vector spaces
α1 v 1 + α2 v 2 + . . . + αn v n ∈ V.
Example 1.2.2. Let V = R2 , let v 1 = (0, 1), v 2 = (1, −1) and v 3 = (2, 2)
be vectors. Let α1 = 1, α2 = 2, α3 = 0 be scalars. The linear combination
of v 1 , v 2 , v 3 with coefficients α1 , α2 , α3 is the vector
Example 1.2.4.
2Z := {n ∈ Z : 2 divides n}
26
Andrea Ferraguti Chapter 1: Vector spaces
first verify that (W, +) is a subgroup of (V, +). If (x, x), (y, y) ∈ W ,
then (x, x) + (y, y) = (x + y, x + y), that is an element of W since
its two entries coincide. Hence + is an operation on W . Clearly
since + is associative on V it is also associative on W . The neutral
element for the sum on V is (0, 0), that belongs to W . Hence it is
the neutral element for the sum on W . Finally, the inverse of an
element (x, x) ∈ W is (−x, −x), that belongs again to W .
Next, λ(x, x) = (λx, λx) ∈ K 2 for every λ ∈ K and (x, x) ∈ W .
Hence the operation K × V → V given by (λ, (x, y)) 7→ (λx, λy)
restricts to an operation on W that makes the latter into a K-vector
space.
On the other hand, the subset U = {(x, 1) : x ∈ K} ⊆ V is not a
vector subspace of V , since (U, +) is not a subgroup of (V, +): in
fact, it has no neutral element.
α1 v 1 + α2 v 2 ∈ W.
27
Andrea Ferraguti Chapter 1: Vector spaces
α1 v 1 + . . . + αn v n ∈ W.
α1 v 1 + . . . + αn v n = w + αn v n ,
28
Andrea Ferraguti Chapter 1: Vector spaces
W = {(x, y, z) ∈ V : x + y + z = 0}.
we get that
W = {p(x) ∈ V : p(0) = 0}
29
Andrea Ferraguti Chapter 1: Vector spaces
In other words, the span of A is the set of all linear combinations of all
elements of A.
30
Andrea Ferraguti Chapter 1: Vector spaces
Proof. In order to prove that two sets are equal, we need to prove that
each is contained in the other one.
So first, suppose that v ∈ hAi. We need to show that P if W is a
subspace that contains A, then v ∈ W . By definition, v = ni=1 αi v i for
some v 1 , . . . , v n ∈ A and α1 , . . . , αn ∈ K. If A ⊆ W , then v 1 , . . . , v n ∈ W
and since subspaces are closed with respect to linear combinations (see
Corollary 1.2.7), it follows that v ∈ W .
Conversely, let v be a vector belonging to every subspace of V that
contains A. Since hAi is a vector subspace of V by Theorem 1.2.11, and
it contains A by Remark 1.2.10, it follows that v ∈ hAi.
What Proposition 1.2.12 says is that the span of a subset A ⊆ V is the smallest
vector subspace of V that contains A.
31
Andrea Ferraguti Chapter 1: Vector spaces
Example 1.2.15.
32
Andrea Ferraguti Chapter 1: Vector spaces
U := {v 0 + w : w ∈ W }
1 , . . . , w n i.
so that u is an element of hv 0 , wP
Conversely, let u = α0 v 0 + ni=1 αi wi ∈ hv 0 , w1 , . . . , wn i. Notice that
v 0 ∈ U , because we can write v 0 as v 0 + 0, that is in U by definition.
Moreover, for every w ∈ W we have −w ∈ W and hence v 0 − w ∈ W .
Therefore, for every w ∈ W we have that:
w = v 0 − (v 0 − w) ∈ hU i,
33
Andrea Ferraguti Chapter 1: Vector spaces
α1 v 1 + α2 v 2 + . . . + αn v n = 0.
Remark 1.3.3.
34
Andrea Ferraguti Chapter 1: Vector spaces
Proof. 1. We have
0 · v 1 + 0 · v 2 + . . . + 0 · v i−1 + 1 · v i + 0 · v i+1 + . . . + 0 · v n = 0,
and hence there is a linear combination of the vectors that gives 0 without
all coefficients being 0.
2. By hypothesis, we have 1 · v i + (−α)v j = 0. Therefore,
n
X
1 · v i + (−α)v j + 0 · vk = 0
k=1
k6=i,j
35
Andrea Ferraguti Chapter 1: Vector spaces
Conversely, if
36
Andrea Ferraguti Chapter 1: Vector spaces
Note:-
Bases are ordered, and therefore will be wrapped in round brackets. That
is, if (v 1 , v 2 , v 3 ) is a basis of a vector space V , then (v 2 , v 1 , v 3 ) is still a basis,
since the concepts of linear dependence and of generation do not depend on
the order, but it is a different basis.
We will now prove that every finitely generated vector space has a basis,
and any two bases have the same cardinality. This requires some preliminary
results.
37
Andrea Ferraguti Chapter 1: Vector spaces
38
Andrea Ferraguti Chapter 1: Vector spaces
w 2 = α1 w 1 + α2 v 2 + . . . + αn v n .
39
Andrea Ferraguti Chapter 1: Vector spaces
Example 1.4.8.
dim K n = n.
40
Andrea Ferraguti Chapter 1: Vector spaces
where Eij is the matrix whose entry in row i and column j is 1 and
all other entries are 0. It follows that
dim V = mn.
Definition 1.4.9. In the vector space K n , the canonical basis is the basis
(e1 , . . . , en ) introduced in Example 1.4.8.
Remark 1.4.10. The K-vector spaces Mn×1 (K) and M1×n (K) of n × 1
and 1 × n matrices, respectively, naturally behave like K n (this claim
can be made formal, but it goes beyond thescope of these notes). In
a
1
layman’s terms, a matrix of the form . . . or a1 . . . an can be
an
n
seen as the element (a1 , . . . , an ) of K . This makes also evident the fact
that dim K n = dim Mn×1 (K) = dim M1×n (K) = n, and that a canonical
basis should also exist for the latter two spaces. We call canonical basis
of M1×n (K) the basis (e1 , . . . , en ) where ei = 0 . . . 1 . . . 0 is the
matrix with a 1 in position (1, i) and 0 elsewhere. We call canonical basis
of Mn×1 (K) the basis constituted of the transpose of the vectors of the
canonical basis of M1×n (K).
The following theorem is of the utmost importance.
41
Andrea Ferraguti Chapter 1: Vector spaces
and hence n
X
(αi − βi )v i = 0.
i=1
42
Andrea Ferraguti Chapter 1: Vector spaces
ΦB : V → K n
n
X
v= ai v i 7→ (a1 , . . . , an )
i=1
is a bijection.
Note:-
The function ΦB is the unique function that associates to a vector the n-
tuple of its components with respect to the basis B.
(a) hv 1 , . . . , v n i = V ;
(b) v 1 , . . . , v n are linearly independent;
(c) (v 1 , . . . , v n ) is a basis of V .
43
Andrea Ferraguti Chapter 1: Vector spaces
44
Andrea Ferraguti Chapter 1: Vector spaces
45
Andrea Ferraguti Chapter 1: Vector spaces
U + W = {u + w : u ∈ U, w ∈ W }.
Proof. First, assume that the sum is direct. We want to show that U ∩
W = {0}. Let v ∈ U ∩ W . Since v ∈ U , a way of expressing v as a sum
of a vector of U and one of W is v = v + 0. On the other hand, v ∈ W
as well, and hence we can also write v = 0 + v, a sum of a vector of U ,
namely 0, and one of W . Since the sum of U and W is direct, there can
be only one way of writing v as a sum of an element of U and one of W ,
and therefore it must be v = 0. That is, U ∩ W = {0}.
Conversely, assume that U ∩ W = {0}. Now let v ∈ U + W , and let
46
Andrea Ferraguti Chapter 1: Vector spaces
1. U ∩ W is finitely generated.
47
Andrea Ferraguti Chapter 1: Vector spaces
α1 u1 + . . . + αm um + β1 w1 + . . . + βn wn = 0.
Then
α1 u1 + . . . + αm um = −β1 w1 − . . . − βn wn . (7)
The left hand side of the above equality is an element of U and it equals the
right hand side that is a vector of W . It follows that α1 u1 + . . . + αm um ∈
U ∩ W = {0}. This means that
α1 u1 + . . . + αm um = 0.
48
Andrea Ferraguti Chapter 1: Vector spaces
The left hand side is a vector of U , and it equals the right hand side that
is a vector of W . Hence the left hand side is a vector of U ∩ W , and
00
therefore it can be expressed PpB . This means that there exist
in the basis
Pn−p
δ1 , . . . , δp ∈ K such that i=1 αi ui = i=1 δi v i . Substituting this in (8),
we get:
β1 = . . . = βn−p = γ1 + δ1 = . . . = γp + δp = 0.
α1 u1 + . . . + αm−p um−p + γ1 v 1 + . . . + γp v p = 0.
α1 = . . . = αm−p = γ1 = . . . = γp = 0.
All in all, we have proved that all the αi ’s, all the βi ’s and all the γi ’s are
0, i.e. that the vectors of C are linearly independent.
49
Andrea Ferraguti Chapter 1: Vector spaces
α1 v 1 + . . . + αm v m − βm+1 v m+1 − . . . − βn v n = 0,
Remark 1.5.8. The proof of Theorem 1.5.7 shows clearly that a subspace
does not have just one direct complement. For example, if V = R2 and
W = h(1, 0)i, then W has infinitely many direct complements. In fact, if
v ∈ V \ W then W 0 = hvi is a direct complement for W : by construction
W ∩ W 0 = {0} and dim(W ⊕ W 0 ) = 2 by Grassmann formula. It must
then be W ⊕ W 0 = V , since the unique subspace of V of dimension 2 is
V itself (see Corollary 1.4.15).
50
Andrea Ferraguti Chapter 2: Determinant and rank
1 2 4
Example 2.1.2. Let A = 0 1 2 ∈ M3 (R).
−1 6 7
1
2 4
Erasing the middle row, we obtain the submatrix .
−1 6 7
4
Erasing the first two columns, we obtain the submatrix 2.
7
Erasingthefirst and third column and the third row, we obtain the
2
submatrix .
1
51
Andrea Ferraguti Chapter 2: Determinant and rank
Example 2.1.4.
a11 a12
• If n = 2 then A = . Definition 2.1.3 yields:
a21 a22
det A = a11 (a22 a33 −a23 a32 )−a12 (a21 a33 −a23 a31 )+a13 (a21 a32 −a22 a31 ).
where A`m is the minor obtained by erasing the `-th row and the m-th
column from A.
In other words, Laplace theorem says that the determinant can be com-
52
Andrea Ferraguti Chapter 2: Determinant and rank
1 2 0
Example 2.1.7. Let A = −1 0 1 ∈ M3 (R). The determinant of A
−1 1 1
can be computed by its definition, namely using the first row:
0 1 −1 1 −1 0
det A = 1 · det − 2 · det + 0 · det =
1 1 −1 1 −1 1
53
Andrea Ferraguti Chapter 2: Determinant and rank
R1
R2
...
5. If A = , where R1 , . . . , Rn are the rows of A, then:
Rk + Rk0
...
Rn
R1 R1
R2 R2
... ...
det A = det + det .
Rk0
Rk
... ...
Rn Rn
54
Andrea Ferraguti Chapter 2: Determinant and rank
Thanks to point 2., it suffices to prove all remaining points for rows,
since the same statement for columns follows by considering the transpose
matrix.
3. Let A = (aij )i,j=1,...,n and suppose A0 = (bij )i,j=1,...,n is obtained by
swapping the i-th row with the (i + 1)-th row of A. By Laplace theorem,
n
X
det A = (−1)i+j aij Aij
j=1
by computing the determinant using the i-th row. On the other hand,
n
X
0
det A = (−1)i+1+j b(i+1)j A0(i+1)j
j=1
4. By induction
on n. If n = 2 and two rows are equal, we necessarily
a b
have A = , so that det A = ab − ab = 0. Now suppose that the
a b
claim is true for a square matrix of size n − 1 and consider A ∈ Mn (K),
with n ≥ 3. Suppose that two rows of A coincide, say Ri = Rj and fix
another row Rk , with k 6= i, j. By Laplace theorem,
n
X
det A = (−1)k+` ak` Ak` ,
`=1
where Ak` is the minor of A obtained by erasing the k-th row and the `-th
column. Clearly, when we erase the k-th row and the `-th column from
A, the submatrix we obtain has two equal rows, because A had them and
55
Andrea Ferraguti Chapter 2: Determinant and rank
we erased a different row. This means that Ak` = 0 for every `, by the
inductive hypothesis. It follows that det A = 0.
5. Let Rk = (ak1 , . . . , akn ) and Rk0 = (a0k1 , . . . , a0kn ). Then by Laplace
theorem using the k-th row we get:
n
X
det A = (−1)k+j (akj + a0kj )Akj =
j=1
n
X n
X
= (−1)k+j akj Akj + (−1)k+j a0kj Akj ,
j=1 j=1
and by Laplace
theorem these twoterms are exactly the determinants of
R1 R1
R2 R2
... ...
the matrices
and
0 .
Rk Rk
... ...
Rn Rn
6. Let A = (aij )i,j=1,...,n and suppose A0 is obtained by A by multiplying
the k-th row by a constant λ ∈ K. By Laplace theorem we get:
n
X n
X
0 k+j
det A = (−1) (λakj )Akj =λ (−1)k+j akj Akj = λ det A.
j=1 j=1
7. Let R1 , . . . , Rn beP
the rows of A and suppose that the k-th row Rk
of A can be written as i6=k λi Ri , for some λi ∈ K. By points 5. and 6.
56
Andrea Ferraguti Chapter 2: Determinant and rank
we get that
R1
...
Ri
X
det A = λi det ,
...
i=1,...,n
i6=k
Ri
...
Rn
and each of the determinants in the right hand side is 0 by point 4.
8. Suppose A0 is obtained from A by adding to the k-th row Rk a
linearPcombination of the other rows, so that the k-th row of A0 equals
Rk + i=1,...,n λi Ri . By point 5. we have
i6=k
R1 R1
R2 R2
... ...
0
det A = det
+ det
P
.
Rk λi Ri
... ...
Rn Rn
Now the first term in the right hand side is exactly det A and the second
term is 0 by point 7.
57
Andrea Ferraguti Chapter 2: Determinant and rank
a a12 . . . a1n
11
0 a22 . . . a2n
size n − 1 and let A =
be upper triangular of size
. . . ... ... ...
0 . . . 0 ann
n. By Laplace theorem, using the first column, we get:
a a . . . a2n
22 23
0 a32 . . . a3n
det A = a11 det
.
. . . . . . . . . . . .
0 . . . 0 ann
In the right hand side of the equality we are computing the determinant
of an upper triangular matrix of size n − 1 and hence by the inductive
hypothesis its determinant is a22 a33 . . . ann . It follows that
58
Andrea Ferraguti Chapter 2: Determinant and rank
Pn 0
Proposition 2.2.2. Let v = i=1 ai v i and let E = (a1 , . . . , an ). Then
t
A · E is the vector of the components of v with respect to the basis B.
Pn Pn
Proof. Let v = j=1 aj v 0j . Now substitute v 0j = i=1 λij v i so that
n n
!
X X
v= aj λij vi
i=1 j=1
Therefore once we have the change of basis matrix, in order to transform the
expression of a vector in basis B 0 into the expression in basis B we just have
to multiply the column vector of components by the matrix A.
59
Andrea Ferraguti Chapter 2: Determinant and rank
is a basis of V , and so is
(2, 0, 0) = −2 · (−1, 0, 0)
Now if we take any vector of V and we write it with respect to the basis
B 0 , such as for example (4, 4, 1) = 1 · (2, 0, 0) + 1 · (0, 1, 1) + 1 · (2, 3, 0), in
order to find its components with respect to B we just need to compute
0 1 3 1 4
0 0 −3 · 1 = −3 .
−2 1 −2 1 −3
In fact,
(4, 4, 1) = 4 · (1, 1, 1) − 3 · (1, 0, 1) − 3 · (−1, 0, 0).
A0 A · t (a1 , . . . , an ) = t (a1 , . . . , an ),
60
Andrea Ferraguti Chapter 2: Determinant and rank
2.3 Rank
Definition 2.3.1. Let A ∈ Mm×n (K) be a matrix. If B is a square sub-
matrix of A, the size of B is the number of rows (or columns) of B.
The rank of A is the largest size of a square submatrix of A with
nonzero determinant. If all square submatrices of A have determinant 0,
then we say that A has rank 0. The rank of A is denoted by rk(A).
Remark 2.3.2.
1. A has rank 0 if and only if all of its entries are 0. In fact, if all of
its entries are 0 then clearly there are no square submatrices with
nonzero determinant, and conversely if all square submatrices of A
have determinant 0 then in particular all submatrices of size 1 have
determinant 0. But a square submatrix of size 1 is just an entry of
A.
61
Andrea Ferraguti Chapter 2: Determinant and rank
Example 2.3.3.
1 2 3
• Let A = ∈ M2×3 (R). Since A is not the zero matrix,
2 4 6
rk(A) ≥ 1 and of course rk(A) ≤ 2. To decide whether rk(A) = 1
or rk(A) = 2 we need to decide if A has a submatrix of size 2 with
non-zero determinant. The submatrices of size 2 of A are:
1 2 1 3 2 3
, and .
2 4 2 6 4 6
Proposition 2.3.4. Let A ∈ Mm×n (K) and let r ≤ min{m, n}. Suppose
that every square submatrix of A of size r + 1 has determinant 0. Then
rk(A) ≤ r.
62
Andrea Ferraguti Chapter 2: Determinant and rank
where B1j is the submatrix obtained from B by erasing the first row and
the j-th column of B. But now for every j = 1, . . . , r + k the matrix B1j
is a square submatrix of B, and hence of A, of size r + k − 1, and hence
it has determinant 0 by the inductive hypothesis. It follows from (9) that
det B = 0.
Corollary 2.3.5. Let A ∈ Mm×n (K) and suppose that A has a size r
submatrix with non-zero determinant. If all submatrices of size r + 1 have
zero determinant, then rk(A) = r
Example 2.3.7. Let V = R3 and let B = ((1, 1, 1), (1, 0, 1), (0, 0, −1)),
that is a basis of V . Let w1 = (2, 1, 0) and w2 = (−1, 0, 0). We have
that w1 = (1, 1, 1) + (1, 0, 1) + 2(0, 0, −1) and w2 = −(1, 0, 1) − (0, 0, −1).
63
Andrea Ferraguti Chapter 2: Determinant and rank
α1 v 1 + . . . + αk v k = 0. (10)
64
Andrea Ferraguti Chapter 2: Determinant and rank
65
Andrea Ferraguti Chapter 2: Determinant and rank
Since det H 6= 0, this means that A has a square submatrix of size k with
non-zero determinant, and hence rk(A) = k.
Example 2.3.11.
66
Andrea Ferraguti Chapter 2: Determinant and rank
Theorem 2.3.8 says that the vectors are linearly independent if and
only if rk(AB ) = 3. The four submatrices of AB of size 3 are:
2 1 0 2 1 0
−1 −1 −1 , −1 −1 −1 ,
3 −1 −5 0 −1 −2
2 1 0−1 −1 −1
3 −1 −5 , 3 −1 −5 ,
0 −1 −2 0 −1 −2
and they all have determinant 0. Therefore rk(AB ) < 3 and the
three vectors are linearly dependent.
67
Andrea Ferraguti Chapter 2: Determinant and rank
The determinant of AB is
2 −1 0
−1 1 −1 0
det AB = − det 0 −1 1 = −2 det −det =
0 1 −1 1
1 0 1
= 2 + 1 = 3 6= 0,
and therefore (v 1 , v 2 , v 3 , v 4 ) is a basis of V by Corollary 2.3.10.
Let A ∈ Mm×n (K). The rows of A can be thought as elements of K n ,
since they are n-tuples of elements of K. Analogously, the columns of A can
be thought as elements of K m .
68
Andrea Ferraguti Chapter 2: Determinant and rank
Proof. Since R(A) = C(t A) and rk(A) = rk(t A), to prove the theorem it
is enough to prove that dim C(A) = rk(A).
Consider the vector space V = K m . The columns C1 , . . . , Cn of A
can be seen as vectors of K m , as in Example 2.3.12. Now consider the
canonical basis (e1 , . . . , em ) of K m . The component matrix of C1 , . . . , Cm
with respect to the canonical basis is exactly the matrix A. Let s =
dim C(A); the matrix A must then have s columns Ci1 , . . . , Cis that are
linearly independent. Let B be the submatrix of A formed by these s
columns. Now we can apply Theorem 2.3.8 to the vector space V and the
matrix B, that is the m × s component matrix of Ci1 , . . . , Cis with respect
to the canonical basis of K m : since these vectors are linearly independent,
the rank of B must be s. Since B is a submatrix of A, we have rk(A) ≥ s.
Now assume by contradiction that r = rk(A) > s. By definition of
rank, this means that there is a square submatrix of A of size r that has
non-zero determinant. Such submatrix is formed by taking r columns
69
Andrea Ferraguti Chapter 2: Determinant and rank
Corollary 2.3.16.
(a) det A 6= 0;
(b) A is invertible;
(c) rk(A) = n;
(d) the rows (columns) of A are linearly independent, when seen
as vectors in K n ;
(e) the rows (columns) of A are a basis of K n .
70
Andrea Ferraguti Chapter 2: Determinant and rank
1 2 3 4
Example 2.3.18. Let A = 5 −1 6 7 ∈ M3×4 (R). The submatrix
7 −4 3 2
1 2
B = has determinant −11 6= 0, so rk(A) ≥ 2. The 3 × 3
5 −1
71
Andrea Ferraguti Chapter 2: Determinant and rank
72
Andrea Ferraguti Chapter 3: Linear systems
If we let
x b
1 1
x2 b2
X=
and ,
B=
. . . . . .
xn bm
we can write linear system (12) in its matrix form
AX = B, (13)
73
Andrea Ferraguti Chapter 3: Linear systems
x
1
x2
∈ Mn×1 (K) is a solution
Definition 3.1.2. A column vector X =
. . .
xn
to system (12) if AX = B.
A linear system is compatible if it admits at least one solution.
Note:-
Notice that there is a big difference between the matrix form of a linear
system (13) and the identity AX = B for some X ∈ Mn×1 (K). In fact,
the former is a formal way of expressing (12), that is a system of equations,
while the latter is an identity between matrices with coefficients in K.
When the number of variables is low, typically less than 5, instead of
labeling them as x1 , . . . , xn we will label them as x, y, z, t.
Example 3.1.3.
• The system (
2x + 3z = 1
y − z = −1
is a linear system of 2 equations and 3 variables with coefficients in
R. The matrix associated to the system is
2 0 3
A= ,
0 1 −1
x
1
and if we let X = y and B = then the matrix form of
−1
z
the system is AX = B.
74
Andrea Ferraguti Chapter 3: Linear systems
−4
A solution is, for example, the vector 2 , and therefore the
3
−7
system is compatible. Another one is 4 . We will see later on
5
that this system has infinitely many solutions.
Now let A ∈ Mm×n (K), X ∈ Mn×1 (K) and B ∈ Mm×1 (K). The key obser-
vation that allows to relate linear systems to vector spaces is the following:
if
x
1
x2
C1 , . . . , Cn are the columns of A, so that A = (C1 |C2 | . . . |Cn ), and X =
,
. . .
xn
then
AX = x1 C1 + x2 C2 + . . . + xn Cn .
Namely, multiplying A by X means taking the linear combination of the
columns of A with coefficients x1 , . . . , xn .
Note:-
If AX = B is a linear system, we will denote by (A|B) the matrixobtained
2 2
by adjoining the column B to the matrix A. For example, if A =
1 −3
75
Andrea Ferraguti Chapter 3: Linear systems
1 2 2 1
and B = , the matrix (A|B) is .
0 1 −3 0
Theorem
3.1.4 (Rouché-Capelli). Let K be a field, A ∈ Mm×n (K), X =
x
1
x2
and B ∈ Mm×1 (K). The system AX = B is compatible if and
. . .
xn
only if rk(A) = rk(A|B).
Proof. The system is compatible if and only if there exists X ∈ Mn×1 (K)
such that AX = B. By what we observed above, this is equivalent to the
existence of coefficients x1 , . . . , xn ∈ K such that
x1 C1 + x2 C2 + . . . + xn Cn = B,
76
Andrea Ferraguti Chapter 3: Linear systems
Here Bi is the matrix obtained from A by replacing the i-th column with
B.
77
Andrea Ferraguti Chapter 3: Linear systems
b
1
b2
ment. Let B =
. Recall that
. . .
bn
n+1
A11 −A21 . . . (−1) An1
n+2
−A12
1 A22 . . . (−1) An2
A−1 = ,
det(A)
... ... ... ...
(−1)1+n A1n (−1)2+n A2n ... Ann
A moment
Pn of reflection, using Laplace theorem on the j-th column, shows
i+j
that b
i=1 i (−1) A ij is exactly the determinant of the matrix Bj =
(C1 | . . . |Cj−1 |B|Cj+1 | . . . |Cn ), where C1 , . . . , Cn are the columns of A.
1 1 −1 x 2
Example 3.1.6. Let A = 1 −1 1 , X = y and B = 0,
0 0 3 z 1
where all coefficients are real. Using Laplace theorem on the last row of
A we see that det A = −6 6= 0, so by Theorem 3.1.5 the system has a
78
Andrea Ferraguti Chapter 3: Linear systems
Example 3.2.1.
1 2 1
• Let A = ∈ M2 (R). The system AX = has a unique
3 4 1
solution, by Theorem 3.1.5.
• Let A = 1 2 ∈ M1×2 (R). The system AX = 0 has infinitely
−2x
many solutions, since all vectors of the form , with x ∈ R,
x
are solutions.
1 1
• Let A = ∈ M2×1 (R). The system AX = has no solu-
2 1
tions, since if x ∈ R = M1 (R) is a solution, then the two equations
x = 1 and 2x = 1 should hold true at the same time.
79
Andrea Ferraguti Chapter 3: Linear systems
so that αX + βY ∈ ker A.
Proof. Start by noticing that hei1 , . . . , eip i coincides with the set of n × 1
matrices (ai1 )i=1...,n ∈ Mn×1 (K) with the property that aj1 = 0 for every
80
Andrea Ferraguti Chapter 3: Linear systems
j∈/ {i1 , . . . , ip }. It follows that the set {AX : X ∈ hei1 , . . . , eip i} coincides
with the set of all linearPcombinations of Ci1 , . . . , CipP : if α1 , . . . , αp ∈ K,
the linear combination j=1 αj Cij corresponds to A( pj=1 αj eij ).
p
The condition ker A ∩ hei1 , . . . , eip i = {0} means that no non-zero vec-
tors in hei1 , . . . , eip i are solutions to the system AX = 0. By what we
said above, this means precisely that no linear combination of Ci1 , . . . , Cip
with not all coefficients being zero gives the zero vector; in other words,
Ci1 , . . . , Cip are linearly independent.
By Lemma 3.2.6, it follows that the columns Ci1 , . . . , Cin−p are linearly
independent. By Theorem 2.3.15, this implies that r ≥ n − p, i.e. p ≥
n − r.
81
Andrea Ferraguti Chapter 3: Linear systems
A(Y − X) = B − B = 0.
This means exactly that Y − X ∈ ker A, or, in other words, that there
exists Z ∈ ker A such that Y − X = Z. Hence Y = X + Z, that is, Y ∈ S.
Conversely, let Y ∈ S, so that there exists Z ∈ ker A such that Y =
X + Z. Then
AY = A(X + Z) = AX + AZ = B + 0 = B,
so that Y ∈ S.
Proposition 3.2.9 completely describes all solutions of a compatible linear
system AX = B: once we find one solution X of the system, all solutions are
of the form X + Z, where Z is any element of the kernel of A.
82
Andrea Ferraguti Chapter 3: Linear systems
1. rk(A0 |B 0 ) = r.
83
Andrea Ferraguti Chapter 3: Linear systems
r
X r
X
λj Rij = Rk and λ j bi j = bk . (16)
j=1 j=1
Since this holds true for every k ∈ {1, . . . , m}, we have by (14) that X is
a solution to AX = B.
What Lemma 3.3.1 is saying is that once we have a compatible linear
system, in order to solve it we can erase equations that are linearly dependent
from the others.
84
Andrea Ferraguti Chapter 3: Linear systems
1 1 0 1
1 −2 2 −1
Example 3.3.2. Let K = R. Let A = and B =
2 −1 2
0
1 4 −2 3
1
0
. Let R1 , . . . , R4 be the rows of A and R e1 , . . . , R
e4 be the rows of
1
2
(A|B). Since R1 , R3 are linearly independent, while R2 = R3 − R1 and
R4 = 3R1 − R3 , we have rk(A) = 2. Moreover, R e2 = R e3 − R
e1 and
R
e4 = 3Re1 − Re3 , so that rk(A|B) = 2 and the system is compatible. By
Lemma 3.3.1, in order to solve the system we can disregard the second
and the fourth equation, i.e. it is enough to solve the system:
x
1 1 0 1 y 1
= .
2 −1 2 0 z
1
t
85
Andrea Ferraguti Chapter 3: Linear systems
where the coefficients cij and d` are elements of k. This means that to ob-
tain all solutions to our linear system, we have to let all variables xk with
k ∈/ {i1 , . . . , im } vary over K, and the values of the remaining variables are
determined by the values of the former ones.
Note:-
Formally, solutions of a linear system AX = B are vectors of Mn×1 (K),
so they should be written as column vectors. However, this is annoying in
practice, so we will often write solutions as vectors in K n , namely as row
vectors.
86
Andrea Ferraguti Chapter 3: Linear systems
1 1 1−t
y = − det = − 1 (2t − 2z − 1).
3 2 1 − 2z 3
Notice how this has exactly the shape predicted by Proposition 3.2.9: a
specific solution of the system is (2/3, 1/3, 0, 0), and the kernel of the
87
Andrea Ferraguti Chapter 3: Linear systems
Note:-
One is not obliged to use Theorem 3.1.5 to solve the system
(
x+y =1−t
.
2x − y = 1 − 2z
For example, here one can also notice that adding up the two equation the
relation
3x = 2 − t − 2z
must hold, so that x = 31 (2 − t − 2z), and substituting this for x in any of
the two equations gives the value of y.
Moreover, one can choose any two linearly independent columns from the
matrix A. For example, a smarter choice here would be to choose the third
and the fourth column, so that the system to be solved becomes:
(
t=1−x−y
,
2z = 1 − 2x + y
This is exactly the same set we found in Example 3.3.3, but written in a
different way.
We end the section by illustrating another way of solving a system in the
form (17). This is called Gauss elimination method, and it is based on the
following observations.
88
Andrea Ferraguti Chapter 3: Linear systems
namely the m × m matrix that differs from the identity matrix just by
the (i, j)-th entry, that is equal to c. Then U is invertible, since it is
lower triangular and the entries on the diagonal are all 1, and U A is an
m × n matrix that coincides with A, except for the fact that the i-throw
is replaced by cRj + Ri .
89
Andrea Ferraguti Chapter 3: Linear systems
That is, the last equation directly gives us the value of xm . Substituting this
into the m − 1-th equation we immediately get the value of xm−1 . Substituting
these values into the (m − 2)-th equation we get the value of xm−2 , and so on.
Notice that the fact that det A 6= 0 is crucial, since it ensures that a0ii 6=
0 for every i = 1, . . . , m. Let us show how this algorithm works with two
examples.
90
Andrea Ferraguti Chapter 3: Linear systems
matrices:
3 2 1 1
A1 = 0 −5/3 2/3 , B1 = −1/3 .
0 −2/3 −4/3 5/3
Next, if R10 , R20 , R30 are the rows of A1 , we replace R30 with R30 − 2/5R20 ,
and we do the same on B1 . We end up with:
3 2 1 1
A2 = 0 −5/3 2/3 , B1 = −1/3 .
0 0 −24/15 27/15
91
Andrea Ferraguti Chapter 3: Linear systems
In order to bring A00 to upper triangular form, the first step is to replace
its second row R2 with R2 − R1 , and to perform the same operation on
92
Andrea Ferraguti Chapter 3: Linear systems
B 00 . This yields:
1 1 0 1 − 2x3
A001 = 0 −1 1 and B100 = −1 + x3 − x5 .
0 1 1 2 − x3 − x5
Now the second and last step is to replace the third row R3 of A001 with
R3 + R2 , and do the same on B100 . We get:
1 1 0 1 − 2x3
00 00
A1 = 0 −1 1 and B1 = −1 + x3 − x5 .
0 0 2 1 − 2x5
System (21) is easy to solve: the third equation tells us that x4 = 1/2 −
x5 ; substituting into the second one we get that x2 = 3/2 − x3 , and
substituting the latter into the first tells us that x1 = −1/2 − x3 . This
means that the set of solutions to (21), and thus to (19), is
Once again, we can see how the set S matches the prediction of Propo-
sition 3.2.9. A specific solution to the system is (−1/2, 3/2, 0, 1/2, 0),
while
ker A = h(−1, −1, 1, 0, 0), (0, 0, 0, −1, 1)i,
that is a two-dimensional subspace of R5 , so that the set S can be written
as
A = {(−1/2, 3/2, 0, 1/2, 0) + Z : Z ∈ ker A}.
93
Andrea Ferraguti Chapter 4: Scalar products and orthogonality
1. for every u, v, w ∈ V ,
(u + v) ∗ w = u ∗ w + v ∗ w;
2. for every u, v, w ∈ V ,
u ∗ (v + w) = u ∗ v + u ∗ w;
v ∗ w = w ∗ v,
Example 4.1.2.
∗: V × V → C
((x1 , x2 , x3 ), (y1 , y2 , y3 )) 7→ x1 y1 + x1 y2 + x2 y2 + x3 y3
is a bilinear form. You can check, as an exercise, that properties
1., 2., 3. are satisfied. However, this is not a scalar product, since
for example (1, 1, 0) ∗ (1, 0, 0) = 1 while (1, 0, 0) ∗ (1, 1, 0) = 2.
94
Andrea Ferraguti Chapter 4: Scalar products and orthogonality
∗: V × V → R
((x1 , x2 ), (y1 , y2 )) 7→ x1 y1 + x1 y2 + x2 y1 + x2 y2
is a scalar product.
Remark 4.1.4.
(x1 , x2 ) ∗ (y1 , y2 ) = x1 y1 − x2 y1 − x1 y2 + x2 y2 .
95
Andrea Ferraguti Chapter 4: Scalar products and orthogonality
1. A⊥ is a vector subspace of V .
2. If A ⊆ B ⊆ V , then B ⊥ ⊆ A⊥ .
3. A⊥ = hAi⊥ .
⊥
4. hAi ⊆ (A⊥ ) .
(α1 v 1 + α2 v 2 ) ∗ w = α1 (v 1 ∗ w) + α2 (v 2 ∗ w) = 0 + 0 = 0,
so that α1 v 1 + α2 v 2 ∈ A⊥ .
2. If v ∈ B ⊥ , then v ∗ w = 0 for every w ∈ B. Since A ⊆ B, it follows
in particular that v ∗ w = 0 for every w ∈ A, and therefore v ∈ A⊥ .
3. Since A ⊆ hAi, by 2. it follows that hAi⊥ ⊆ A⊥ . Conversely,
suppose that v ∈ A⊥ . Let w ∈ hAi. Then, by definition Pn of span, there
exist α1 , . . . , αn ∈ K and v 1 , . . . , v n ∈ A such that w = i=1 αi v i . Then:
n
! n
X X
v∗w =v∗ αi v i = αi (v ∗ v i ) = 0
i=1 i=1
96
Andrea Ferraguti Chapter 4: Scalar products and orthogonality
2. v ∗ v = 0 if and only if v = 0.
Rn × Rn → R
n
X
((x1 , . . . , xn ), (y1 , . . . , yn )) 7→ xi yi .
i=1
R2 × R2 7→ R
kαvk = |α|kvk.
97
Andrea Ferraguti Chapter 4: Scalar products and orthogonality
Proof. p p
kαvk = (αv) · (αv) = α2 (v · v) = |α|kvk.
|u · v| ≤ kukkvk,
98
Andrea Ferraguti Chapter 4: Scalar products and orthogonality
99
Andrea Ferraguti Chapter 4: Scalar products and orthogonality
100
Andrea Ferraguti Chapter 4: Scalar products and orthogonality
Example 4.2.11.
v 01 = v 1 = (1, 1, 1).
Next, we let
v 2 · v 01 0 2
v 02 = v 2 − 0 2 1
v = (1, 0, 1) − (1, 1, 1) = (1/3, −2/3, 1/3).
kv 1 k 3
Finally, we let
v 3 · v 02 0 v 3 · v 01 0
v 03 = v 3 − v − v =
kv 02 k2 2 kv 01 k2 1
2/3 2
= (0, 0, 2) − (1/3, −2/3, 1/3) − (1, 1, 1) = (−1, 0, 1).
2/3 3
The sequence ((1, 1, 1), (1/3, −2/3, 1/3), (−1, 0, 1)) is then an or-
thogonal basis of V . In order to obtain an orthonormal basis, we
just need to divide every vector by its norm. This yields:
√ ! !
1 1 1 1 2 1 1 1
B0 = √ , √ , √ , √ , −√ , √ , − √ , 0, √ .
3 3 3 6 3 6 2 2
R3 × R3 → R
(x1 , x2 , x3 ) · (y1 , y2 , y3 )) =
= 2x1 y1 + x1 y2 + x2 y1 + 4x2 y2 + x2 y3 + x3 y2 + 2x3 y3 .
The reader can verify as an exercise that this is indeed a posi-
tive definite scalar product. Now consider the canonical basis B =
101
Andrea Ferraguti Chapter 4: Scalar products and orthogonality
((1, 0, 0), (0, 1, 0), (0, 0, 1)). This is of course an orthonormal basis
with respect to the standard scalar product on V , but it is no longer
orthonormal (and neither orthogonal) with respect to the above
scalar product. Hence, we can apply Gram-Schmidt algorithm and
obtain an orthonormal one. First, we set
Next, we let
v 01 = v 1 = (1, 0, 0).
Notice that kv 01 k2 = 2. Next, we let:
v 2 · v 01 0 1
v 02 = v 2 − 0 2 1
v = (0, 1, 0) − (1, 0, 0) = (−1/2, 1, 0).
kv 1 k 2
v 3 · v 02 0 v 3 · v 01 0
v 03 = v 3 − v − v =
kv 02 k2 2 kv 01 k2 1
1 0
(0, 0, 1) − (−1/2, 1, 0) − (1, 0, 0) = (1/7, −2/7, 1),
7/2 2
so that kv 03 k2 = 12/7. Hence an orthonormal basis for this scalar
product is:
1 1 2 1 2 7
√ , 0, 0 , − √ , √ , 0 , √ , − √ , √ .
2 14 14 84 84 84
102
Andrea Ferraguti Chapter 4: Scalar products and orthogonality
Proof. First, we need to show that hAi ∩ A⊥ = {0}. This is easy: first
recall that, by Proposition 4.1.6, A⊥ = hAi⊥ . Hence if v ∈ hAi ∩ A⊥ =
hAi ∩ hAi⊥ then we must have v · v = 0. But v · v = kvk2 , and hence
kvk2 = 0. Since the scalar product is positive definite, this implies that
v = 0.
Next, let m = dim A. To conclude the proof, we need to sho that
dim A⊥ = n − m. By Grassmann formula, we have:
dim(hAi ⊕ A⊥ ) = m + dim A⊥ ,
dim A⊥ ≤ n − m.
dim A⊥ ≥ n − m,
103
Andrea Ferraguti Chapter 4: Scalar products and orthogonality
104
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
The eigenvalues of the matrix A are the roots of the characteristic poly-
nomial.
Example 5.1.3.
1 2
• Let A = ∈ M2 (R). Then
1 0
1−x 2
pA (x) = det = x2 − x − 2.
1 −x
1 0 −1
• Let A = 0 1 −1 ∈ M3 (R). Then
−1 −1 0
1−x 0 −1
3 2
pA (x) = det 0 1 − x −1 = x − 3x − x + 2
−1 −1 −x
105
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
AX = λX.
Vλ = ker(A − λIn ).
Note:-
Technically, an eigenspace is a subspace of Mn×1 (K), so eigenvectors will be
column vectors. However, since practically it is easier to write rows instead
of columns, we will think of eigenspace as subspaces of K n , i.e. we will write
eigenvectors n-tuples of elements of K.
λ1 0 ... 0
0 λ2 . . . 0
Remark 5.1.6. If D = ∈ Mn (K) is a diagonal matrix
. . . . . . . . . . . .
0 . . . 0 λn
then
λ −x 0 ... 0
1
n
λ2 − x . . .
0 0 Y
pD (x) = det(D−xIn ) = det
=
(λi −x).
... ... ... . . . i=1
0 ... 0 λn − x
106
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
AP = P D.
107
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
the zero vector since P is invertible). Moreover, since det P 6= 0 then the
columns of P are a basis of K n .
Conversely, let P1 , . . . , Pn be a basis of K n consisting of eigenvectors
of A. Let P = (P1 | . . . |Pn ). Since AP = (AP1 | . . . |APn ) and the Pi ’s are
eigenvectors, there exist λ1 , . . . , λn ∈ K such that APi = λi Pi for every
i = 1, . . . , n. Therefore,
P −1 AP = D,
so that A is diagonalizable.
Remark 5.1.9. The proof of Theorem 5.1.8 shows how to find a diagonal-
izing matrix for a diagonalizable matrix A ∈ Mn (K): it sufficies to find
a basis (P1 , . . . , Pn ) of K n consisting of eigenvectors of A and then set
P = (P1 |P2 | . . . |Pn ). The matrix P is a diagonalizing matrix for A.
Example 5.1.10.
1 2
• Let A = ∈ M2 (R). The characteristic polynomial of A is:
2 1
1−x 2
PA (x) = det = (1 − x)2 − 4,
2 1−x
108
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
{(α, −α) : α ∈ R} ⊆ R2 .
Next,
−2 2
V3 = ker(A − 3I2 ) = ker = {(α, α) : α ∈ R}.
2 −2
1 1
• Let A = ∈ M2 (R). We have pA (x) = (1 − x)2 , so that
0 1
the unique eigenvalue
ofA is λ = 1. The relative eigenspace is
0 1
ker(A − I2 ) = ker , that is a 1-dimensional subspace of R2 .
0 0
2
Therefore, R cannot have a basis consisting of eigenvectors of A; it
follows that A is not diagonalizable.
109
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
Next,
0 −1
0
V1 = ker(A − I2 ) = ker 0 0 −1 ,
−1 −1 −1
so that
V1 = {(α, −α, 0) : α ∈ R} = h(1, −1, 0)i.
110
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
Finally,
−1 0 −1
V2 = ker(A − 2I2 ) = ker 0 −1 −1
−1 −1 −2
so that
V2 = {(α, α, −α) : α ∈ R} = h(1, 1, −1)i.
3
The sequence ((1, 1, 2), (1, −1, 0), (1, 1, −1)) is a basis
of R , and
1 1 1
therefore the matrix A is diagonalizable. If P = 1 −1 1
2 0 −1
then
−1 0 0
P −1 AP = 0 1 0 .
0 0 2
111
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
X = X1 + X2 + . . . + Xm = Y1 + Y2 + . . . + Ym
112
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
we get:
(X1 − Y1 ) + (X2 − Y2 ) + . . . + (Xm − Ym ) = 0. (22)
Multiplying by A on the left both sides of (22) we get:
(λ1 −λm )X1 +. . .+(λm−1 −λm )Xm−1 = (λ1 −λm )Y1 +. . .+(λm−1 −λm )Ym−1 .
113
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
n = gλ1 + . . . + gλm .
On the other hand, gλi ≤ aλi for every i by Theorem 5.1.13, and of course
aλ1 +. . .+aλm ≤ n by the fundamental theorem of algebra 0.5.3. It follows
that:
n = gλ1 + . . . + gλm ≤ aλ1 + . . . + aλm ≤ n,
and hence equality must hold everywhere. In particular, aλ1 + . . . + aλm =
n, so that λ1 , . . . , λm are all the eigenvalues of A, and so they all belong
to K, and gλi = aλi for every i, so that all eigenvalues are regular.
Conversely, suppose that all eigenvalues of A belong to K and they
are all regular. Let such eigenvalues be λ1 , . . . , λm . Since aλi = gλi for
every i, we have that gλ1 + . . . + gλm = n because, as we explained above,
aλ1 + . . . + aλm = n by the fundamental theorem of algebra. On the other
hand,
dim(Vλ1 + . . . + Vλm ) = dim(Vλ1 ⊕ . . . ⊕ Vλm )
by Proposition 5.1.16. Altogether, these observations imply that
V λ1 ⊕ . . . ⊕ V λm = K n .
Example 5.1.18.
is not diagonalizable. In fact, such matrix does not satisfy the hy-
114
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
X = (x1 , . . . , xn ).
115
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
AX = AX = λX = λ · X, (26)
using the fact that A = A since A has real entries. This means that
X is an eigenvector of A relative to the eigenvalue λ. Using (26) together
with the fact that A = t A we get:
(λ − λ)(t X · X) = 0.
Note:-
The standard scalar product Cn × Cn → C defined above is not positive
definite; in fact it makes no sense to talk about positive complex numbers.
Moreover, it is not even true that (x1 , . . . , xn ) • (x1 , . . . , xn ) = 0 if and only
if (x1 , . . . , xn ) = 0. For example, (1, i) • (1, i) = 1 − 1 = 0.
116
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
√ √
1/ 2 1/ 2
Example 5.2.5. The matrix A = √ √ ∈ M2 (R) is orthogo-
1/ 2 −1/ 2
√ √ √ √
nal, since ((1/ 2, 1/ 2), (1/ 2, −1/ 2)) is an orthonormal basis of R2 .
1. Ri • Rj = 0 for every i 6= j;
2. Ri • Ri = 1 for every i;
3. Ci • Cj = 0 for every i 6= j;
117
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
4. Ci • Ci = 1 for every i.
On the other hand, t A = A−1 if and only if the following conditions hold:
(a) A · t A = In ;
(b) t A · A = In .
118
Andrea Ferraguti Chapter 5: Eigenspaces and diagonalization
119
Andrea Ferraguti Chapter 6: Affine geometry
One shall think of A as the set of points in the affine space, while elements of V
are vectors that represent the ”difference” between two points. The following
examples are the keys to understanding the concept of affine space.
f (P, Q) := (xQ − xP , yQ − yP ) ∈ V.
120
Andrea Ferraguti Chapter 6: Affine geometry
f (P, R) = (xR − xP , yR − yP ),
Example 6.1.2 is nothing else than the well-known cartesian plane. The
points of the space are given by pairs of real numbers, and for every pair of
points P, Q in the plane there is a vector connecting them. This vector can be
visualized by drawing a straight arrow from P to Q pointing at Q. However,
one needs to be careful using this graphic representation because vectors with
the same length, direction and verse are the same element of the underlying
vector space V . For example, the vector connecting the points (0, 0) and (1, 0)
is the same vector that connects (0, 1) and (1, 1), although graphically we
represent the two vectors as two separate objects.
The function f of example 6.1.2 is nothing else that the function that
associates to a pair of points P, Q the vector of V = R2 that connects them.
Property (i) then just says that given a point P and a vector v, one can
”translate” the point P by v in a unique way, obtaining a new point Q. For
example, if one is given the point P = (1, 1) and the vector v = (−2, 0), the
unique translate of P by v is the point Q = (−1, 1). Finally, property (ii)
simply says that if Q is the translate of P by a vector v and R is the translate
of Q by a vector w then R is the translate of P by the vector v + w.
Example 6.1.2 can of course be generalized to any dimension and any field.
We will see later on that this is, in a way, the only relevant example.
Then the triple (A, f, V ) is an affine space of dimension n over the field K.
Of course when n = 2 and K = R we recover Example 6.1.2. When n = 3
and K = R this is also a well-known object: it is nothing else that the
3-dimensional space, where every point is identified uniquely by a triple
of real numbers, usually referred to as ”coordinates”. Once again, given
two points in the space it is possible to connect them by a unique vector,
represented as a straight arrow starting at the first point and pointing at
121
Andrea Ferraguti Chapter 6: Affine geometry
Definition 6.1.4. The affine space defined in Example 6.1.3 will be de-
noted by An (K).
From now on, when (A, V, f ) is an affine space and P, Q ∈ A, we will denote
−→
by P Q the vector f (P, Q).
Proposition 6.1.5. Let (A, V, f ) be an affine space over a field K, and let
P, Q, R, S ∈ A. The following hold true.
−→ −→
1. P Q = P R if and only if Q = R;
−→
2. P Q = 0 if and only if P = Q;
−→ −→
3. if v = P Q, then −v = QP ;
−→ −→ −→ −→
4. P Q = RS if and only if P R = QS.
−→ −→ −→
Proof. 1. Of course if Q = R then P Q = P R. Conversely, let v := P Q.
By property (i) of the function f , there exists a unique point X ∈ A such
−−→ −→
that P X = v. Therefore that point must be Q, and hence if P R = v then
necessarily R = Q.
−→ −→ −→
2. By property (ii) of the function f , we have P P = P P + P P , and
−→ −→ −→ −→
therefore P P = 0. Conversely, if P Q = 0, then P Q = P P , and therefore
Q = P by point 1.
−→ −→ −→
3. By property (ii) of the function f , we have P P = P Q + QP . By
−→ −→
point 2., this implies that P Q = −QP .
−→ −→ −→ −→ −→
4. By property (ii) of f , we have P Q + QS = P S = P R + RS. The
claim follows immediately.
122
Andrea Ferraguti Chapter 6: Affine geometry
tv : A → A
P 7→ tv (P )
Note:-
The existence of the translate of P by v is granted by property 1. of the
function f associated to an affine space.
In other words, a linear subspace of an affine space is the set of all translates
of a point via vectors that lie in a vector subspace of V .
123
Andrea Ferraguti Chapter 6: Affine geometry
Example 6.2.2. Let A2 (R) be the affine space of Example 6.1.2. Let
W ⊆ V be the subspace generated by the vector (1, 1) and let O =
(0, 0) ∈ A2 (R). The linear subspace [O, W ] is the set of all points in the
plane that are translates of (0, 0) via a multiple of (1, 1). Depicting A2 (R)
as the usual cartesian plane, [O, W ] is nothing else than the bisector of
the first quadrant.
Similarly, let A3 (R) be the affine space of Example 6.1.3 with n = 3
and K = R. Let O = (1, 1, 1) ∈ A3 (R) and let W = h(1, 0, 0), (0, 1, 0)i ⊆
V . Depicting A3 (R) in the usual way, the linear subspace [O, W ] is the
plane passing by (1, 1, 1) and parallel to the xy-plane.
In general, we will see that linear subspaces in An (K) are nothing else
than sets of solutions of linear systems!
124
Andrea Ferraguti Chapter 6: Affine geometry
125
Andrea Ferraguti Chapter 6: Affine geometry
so that Q ∈ [O, W ].
Proposition 6.2.6. Let (A, V, f ) be an affine space over a field K and let
[O, W ] and [O0 , W 0 ] be two linear subspaces.
126
Andrea Ferraguti Chapter 6: Affine geometry
1. [O, W ] ⊆ [O0 , W 0 ];
2. [O0 , W 0 ] ⊆ [O, W ];
3. [O, W ] ∩ [O0 , W 0 ] = ∅.
127
Andrea Ferraguti Chapter 6: Affine geometry
Proof. Let r = [P, hvi] and s = [Q, hwi] for some non-zero vectors v, w ∈
V . By contradiction, suppose that r and s are not parallel. Then w ∈ /
hvi, as otherwise we would have hwi ⊆ hvi. Hence v and w are linearly
independent, and therefore dimhv, wi = 2. Since dim V = 2, it follows
−→
that V = hv, wi. Hence there exist α, β ∈ K such that P Q = αv + βw.
−−→
Let Q0 ∈ s be such that QQ0 = −βw. Then:
−−→0 −→ −−→0
P Q = P Q + QQ = αv + βw − βw = αv.
1. Two lines r, s in (A, V, f ) are skew if and only if they are not coplanar
Proof. 1. First, assume that r, s are skew. Let r = [P, hvi] and s =
[Q, hwi]. Suppose by contradiction that they lie on a plane π = [O, W ],
where W ⊆ V is a subspace of dimension 2. By Proposition 6.2.3, π is an
affine space of dimension 2. Since r ∩ s = ∅, by Proposition 6.3.4 they are
parallel, but this contradicts the hypothesis.
Conversely, assume that there is no plane containing r and s. Assume
by contradiction that r ∩ s 6= ∅. Let P ∈ r ∩ s. Write r = [P, hvi] and
s = [P, hwi] for some non-zero v, w ∈ V , thanks to Proposition 6.2.5. The
space hv, wi is at most 2-dimensional, hence there exists a subspace W ⊆
V such that dim W = 2 and hv, wi ⊆ W . Then the linear subspace [P, W ]
is a plane that clearly contains both lines, contradicting the hypothesis.
128
Andrea Ferraguti Chapter 6: Affine geometry
129
Andrea Ferraguti Chapter 6: Affine geometry
−→
that V = hu, v, wi. Therefore P Q = αu + βv + γw for some α, β, γ ∈ K.
−−→
Let Q0 ∈ π be the point such that QQ0 = −βv − γw. Then
−−→0 −→ −−→0
P Q = P Q + QQ = αu.
π ∩ σ = [P, W ∩ W 0 ],
Definition 6.3.8. Let (A, V, f ) be an affine space. Two lines that intersect
in a single point, a line and a plane that intersect in a single point or two
planes that intersect in a single line are called incident.
130
Andrea Ferraguti Chapter 6: Affine geometry
1. The map
Λ[O,B] : A → K n
P 7→ (x1 , . . . , xn ),
131
Andrea Ferraguti Chapter 6: Affine geometry
−→ P
where (x1 , . . . , xn ) ∈ K n is the only n-tuple such that OP = ni=1 xi v i ,
is a bijection.
Proof. 1. First, suppose that P, Q ∈ A are such that Λ[O,B] (P ) = Λ[O,B] (Q) =
−→ P −→
(x1 , . . . , xn ). This means that OP = ni=1 xi v i = OQ. However, P by defini-
tion of affine space, given the point O ∈ A and the vector ni=1 xi v i ∈ V ,
−−→ P
there exists a unique point X ∈ A such that OX = ni=1 xi v i . This means
that P = X = Q. Therefore Λ[O,B] is injective. Pn
Next, if (x1 , . . . , xn ) ∈ K n , let v = i=1 xi v i ∈ V . Once again by
−→
definition of affine space, there exists a unique P ∈ A such that OP = v.
But then Λ[O,B] (P ) = (x1 , . . . , xn ). This proves that Λ[O,B] is surjective as
well.
2. Since Λ[O,B] (P ) = (x1 , . . . , xn ) and Λ[O,B] (Q) = (y1 , . . . , yn ), the
−→ P −→ P
vector OP is ni=1 xi v i while OQ = ni=1 yi v i . It follows that
n n n
−→ −→ −→ X X X
P Q = P O + OQ = − xi v i + yi v i = (yi − xi )v i ,
i=1 i=1 i=1
and therefore
−→
ΦB (P Q) = (y1 − x1 , . . . , yn − xn ).
Theorem 6.4.2 is of the utmost importance. It shows that every affine space
of dimension n over a field K, once a coordinate system is chosen, ”becomes”
the affine space An (K) described in Example 6.1.3. The bijection Λ[O,B] , that
depends on the choice of a coordinate system, yields a dictionary that allows
to translate all properties of an affine space of dimension n over a field K into
properties of the very concrete object An (K).
132
Andrea Ferraguti Chapter 6: Affine geometry
Note:-
−→
Notice that since OO = 0, we have that Λ[O,B] (O) = (0, . . . , 0). Namely, the
map Λ[O,B] transforms the origin O of a coordinate system into the point
(0, . . . , 0) ∈ K n .
Moreover, ΦB (0) = (0, . . . , 0).
From now on, with a slight abuse of notation, we will write An (K) to
denote the set of points of the latter affine space, although technically An (K)
denotes a triple of objects, according to the definition of affine space.
The next theorem describes how linear subspaces transform under Λ[O,B] .
First, we prove two preliminary lemmas.
so that
On the other hand, αΦB (w1 ) = α(a1 , . . . , an ) and βΦB (w2 ) = β(b1 , . . . , bn ),
and the claim follows easily.
2. First, we show that ΦB (W ) is a subspace of K n . Let w1 , w2 ∈
ΦB (W ), so that there exist u1 , u2 ∈ W such that ΦB (u1 ) = w1 and
ΦB (u2 ) = w2 . Let α, β ∈ K. Then
133
Andrea Ferraguti Chapter 6: Affine geometry
Lemma 6.4.4. Let • be the standard scalar product on K n , i.e. the map
Kn × Kn → K
n
X
((x1 , . . . , xn ), (y1 , . . . , yn )) 7→ xi y i .
i=1
n
Let W be a subspace of K .
1. dim W ⊥ = n − dim W .
2. (W ⊥ )⊥ = W .
134
Andrea Ferraguti Chapter 6: Affine geometry
W ⊥ = ker A. (28)
(w1 • v, . . . , wm • v) = Av,
135
Andrea Ferraguti Chapter 6: Affine geometry
[Λ[O,B] (P ), ΦB (W )].
It follows that
n
!
X
Φ (yi − xi )v i = (y1 − x1 , . . . , yn − xn ) ∈ Φ(W ),
i=1
and since Λ(Q) = Λ(P )+(y1 −x1 , . . . , yn −xn ), we get precisely that Λ(Q)
is a translate of Λ(P ) via a vector of Φ(W ), i.e. that Λ(Q) ∈ [Λ(P ), Φ(W )].
We have thus proved that Λ([P, W ]) ⊆ [Λ(P ), Φ(W )].
Conversely, let (y1 , . . . , yn ) ∈ [Λ(P ), Φ(W )]. Then
136
Andrea Ferraguti Chapter 6: Affine geometry
that n
−→ −→ −→ X
P Q = OQ − OP = ai v i .
i=1
Pn
Now Φ ( i=1 ai v i ) = (a1 , . . . , an ), and since the map Φ isP a bijection and
(a1 , . . . , an ) ∈ Φ(W ) by hypothesis, we must have that ni=1 ai v i ∈ W .
−→
Hence P Q ∈ W , that is, Q is a translate of P via a vector of W , i.e.
Q ∈ [P, W ]. Therefore Λ(Q) = (y1 , . . . , yn ) ∈ Λ([P, W ]). It follows that
[Λ(P ), Φ(W )] ⊆ Λ([P, W ]).
2. First, suppose that S is a linear subspace. Then S = [P, W ], for
some P ∈ K n and W vector subspace of K n . By Lemma 6.4.4, there
exists an (n − m) × n matrix A of rank n − m such that W = ker A. Now
if P = (x1 , . . . , xn ) then S, as a subset of An (K) = K n , is nothing else
that
{(x1 , . . . , xn ) + (z1 , . . . , zn ) : (z1 , . . . , zn ) ∈ ker A}.
x
1
Setting B := A . . . and appealing to Proposition 3.2.9, it follows that
xn
S is the set of solutions of the linear system AX = B.
Conversely, if AX = B is a linear system of n − m equations in n
variables and rk A = n − m, then by Theorem 3.2.7 and Proposition 3.2.9,
if (x1 , . . . , xn ) ∈ K n is a solution of the system then the set S of all
solutions is
Theorem 6.4.5 essentially tells us that we can reduce the study of linear sub-
spaces of affine spaces to the study of solutions of linear systems.
137
Andrea Ferraguti Chapter 6: Affine geometry
Definition 6.5.1. A system of the form (29) that describes a line in An (K)
is called a system of cartesian equations for a line.
a11 x1 + a12 x2 = b1 ,
where the rank condition simply means that (a11 , a12 ) 6= (0, 0).
In A3 (K), a system of cartesian equations for a line has the form:
(
a11 x1 + a12 x2 + a13 x3 = b1
,
a21 x1 + a22 x2 + a23 x3 = b2
or, in other words, that the two vectors (a11 , a12 , a13 ) and (a21 , a22 , a23 )
are linearly independent.
The description of a line ` via a system of the form (29) is implicit, namely,
we don’t have a parametrization for the points of the line. To do this, it is
enough to solve the system. Notice that since rk(A) = n − 1, Theorem 3.2.7
implies that ker A is a 1-dimensional vector subspace of K n , that is therefore
138
Andrea Ferraguti Chapter 6: Affine geometry
(p1 , . . . , pn ) + t(a1 , . . . , an )
We stress the fact that (p1 , . . . , pn ) and (a1 , . . . , an ) are two elements of K n ,
the second one being non-zero, while t is a parameter. The point (p1 , . . . , pn )
is a point lying on the line, corresponding to t = 0. All other points can be
found by letting t vary in K.
Since the above matrix has 2 lines and n columns, and the bottom row is non-
zero, asking that it has rank 1 is equivalent to asking that every 2×2 submatrix
has determinant 0. Again, since the bottom row is non-zero there exists some
i ∈ {1, . . . , n} such that ai 6= 0. Then by Theorem 2.3.17 the condition on
139
Andrea Ferraguti Chapter 6: Affine geometry
the rank is satisfied if and only if every 2 × 2 submatrix containing the i-th
column has determinant 0. Assuming that a1 6= 0, we get:
x1 − p 1 xi − p i
det = 0 for every i = 2, . . . , n,
a1 ai
that clearly has rank n − 1 because if we erase the first column the resulting
square matrix has determinant (−a1 )n−1 6= 0, since we assumed that a1 6=
0.
Example 6.5.4. Consider the line ` ⊆ A3 (R) given by the system of equa-
tions: (
2x + 3y + z = 1
`: .
x+z =3
In order to find a system of parametric equations for `, all we have to do
is to solve the system. Letting
2 3 1
A=
1 0 1
be the matrix associated with the system, we see that it has rank 2, and
140
Andrea Ferraguti Chapter 6: Affine geometry
Notice that the line ` passes through the point (3, −5/3, 0).
Example 6.5.5. Consider the line ` ⊆ A3 (R) given by the system of para-
metric equations:
x = 1 + t
y = −t .
z = 1 + 2t
141
Andrea Ferraguti Chapter 6: Affine geometry
π : a1 x1 + a2 x2 + . . . + an xn = b,
142
Andrea Ferraguti Chapter 6: Affine geometry
that
S is parallel to S 0 ⇐⇒ ker A ⊆ ker A0 or ker A0 ⊆ ker A. (31)
On the other hand, by Theorem 2.3.15 we have that dim R(A) = m and
dim R(B) = p, so by (32) and (33) we get that
It follows that the rows of A are a basis of (ker A)⊥ and those of B are
a basis of (ker B)⊥ .
143
Andrea Ferraguti Chapter 6: Affine geometry
We can now proceed to describe relative positions of lines and planes in A2 (K)
and A3 (K) (as we did in Theorems 6.3.9,6.3.10 and 6.3.11) but using their
equations.
1. ` k `0 if and only if
a b
rk = 1.
0 0
a b
If ` k `0 , then ` = `0 if and only if
a b c
rk = 1.
0 0 0
a b c
Proof. By Remark 6.4.6, lines ` and `0 are associated with the linear
144
Andrea Ferraguti Chapter 6: Affine geometry
and their translation spaces are ker A and ker A0 , respectively. Since
rk(A) = rk(A0 ) = 1, by (31) and Lemma 6.6.1, we have that ` k `0 if
and only if
a b
rk = 1,
0 0
a b
and consequently ` and `0 are incident if and only if
a b
rk = 2.
a0 b 0
1. π k π 0 if and only if
a b c
rk = 1.
0 0 0
a b c
145
Andrea Ferraguti Chapter 6: Affine geometry
Proof. The translation spaces of the planes π, π 0 are ker A, ker A0 , respec-
tively, where
A = a b c and A0 = a0 b0 c0 .
0 0
Since rk(A) = rk(A ) = 1, by (31) and Lemma 6.6.1, it follows that π k π
a b c
if and only if rk = 1.
a0 b 0 c 0
In this case, by Proposition 6.2.8 we have π = π 0 if and only if π ∩ π 0 6=
∅, that is, if and only if the system
(
ax + by + cz = d
a0 x + b0 y + c0 z = d0
a b c
is compatible, and since rk = 1, this happens precisely when
0 0 0
a b c
a b c d
rk = 1.
a0 b 0 c 0 d 0
146
Andrea Ferraguti Chapter 6: Affine geometry
1. ` k π if and only if
a b c
1 1 1
rk a2 b2 c2 = 2.
a3 b 3 c 3
If ` and π are parallel, then ` ⊆ π if and only if
a b1 c 1 d 1
1
rk a2 b2 c2 d2 = 2.
a3 b3 c 3 d 3
147
Andrea Ferraguti Chapter 6: Affine geometry
a b c
1 1 1
is compatible, and since rk a2 b2 c2 = 2 this happens precisely when
a3 b 3 c 3
a b c d
1 1 1 1
rk a2 b2 c2 d2 = 2.
a3 b 3 c 3 d 3
148
Andrea Ferraguti Chapter 6: Affine geometry
149
Andrea Ferraguti Chapter 6: Affine geometry
a b c
1 1 1
a2 b 2 c 2
is compatible, and since rk
= 2 this happens precisely when
a3 b 3 c 3
a4 b 4 c 4
a b c d
1 1 1 1
a2 b 2 c 2 d 2
rk = 2. If ` and `0 are not parallel, then either they
a3 b 3 c 3 d 3
a4 b 4 c 4 d 4
are skew or they are incident. This depends on ` ∩ `0 , that is empty
in the first case and has one point in the second. This is governed by
system (35): the system has no solution when the lines are skew and has
precisely one solution when they are incident. Theorem 3.1.4 allows then
to conclude.
Definition 6.7.1. A pencil of lines in A2 (K) is the set of all lines that
pass through a given point.
An improper pencil of lines in A2 (K) is the set of all lines that are
parallel to a given one.
The next proposition explains how to write down equations for pencils of
lines in A2 (K).
Proposition 6.7.2.
150
Andrea Ferraguti Chapter 6: Affine geometry
ax + by + k = 0.
Proof. 1. First, assume that the line ` passes through P . Then a00 xP +
b00 yP + c00 = 0. This means that the vectors (a00 , b00 , c00 ) and (xP , yP , 1) are
orthogonal with respect to the standard scalar product on K 3 . Hence,
(a00 , b00 , c00 ) ∈ h(xP , yP , 1)i⊥ . Now since r and s both pass through P , the
same reasoning holds true, so that
On the one hand the space h(xP , yP , 1)i⊥ has dimension 2 by Lemma 6.4.4.
Since by hypothesis the lines r and s are distinct, the vectors (a, b, c) and
(a0 , b0 , c0 ) are linearly independent in K 3 . It follows that these two vector
form a basis of h(xP , yP , 1)i⊥ , and therefore there exist λ, µ ∈ K such that
as required.
Conversely, if the equation of ` is λ(ax + by + c) + µ(a0 x + b0 y + c0 ) = 0
then ` passes through P since axP + byP + c = a0 xP + b0 yP + c = 0 by
hypothesis, and therefore
Example 6.7.3. Let (1, 1) ∈ A2 (R). In order to find the pencil of lines
through (1, 1) we first need to find two distinct lines through such point.
For example, we can pick x − 1 = 0 and y − 1 = 0. Next, the equation of
the pencil is simply:
λ(x − 1) + µ(y − 1) = 0.
Remark 6.7.4. One might think that the pencil of lines through a point
(xP , yP ) is y − yP = m(x − xP ), where m ∈ K. However, this equation
misses one of the lines of the pencil, namely the line x = xP . In fact
151
Andrea Ferraguti Chapter 6: Affine geometry
this equation comes from the following simplification: given the correct
equation of the pencil of lines through P , namely λr + µs = 0 for some
lines r, s through P , we can divide everything by µ, since proportional
equations give rise to the same line. This way we obtain an equation that
depends only on one parameter, namely λ/µ, but we miss the equation of
the line of the pencil that corresponds to µ = 0.
Definition 6.7.5. A pencil of planes in A3 (K) is the set of all planes that
contain a given line.
An improper pencil of planes in A3 (K) is the set of all planes that are
parallel to a given one.
Proposition 6.7.6.
ax + by + cz + k = 0.
x = x0 + v1 t
Proof. 1. Let ` : y = y0 + v2 t be a parametric equation for `, with
z = z0 + v3 t
(v1 , v2 , v3 ) 6= (0, 0, 0).
First, assume that ϑ contains `. Then we must have:
152
Andrea Ferraguti Chapter 6: Affine geometry
In turn, these two conditions are equivalent to asking that the vector
(a00 , b00 , c00 , d00 ) is orthogonal, via the standard scalar product on K 4 , to
both (x0 , y0 , z0 , 1) and (v1 , v2 , v3 , 0). In other words, if we let W :=
h(x0 , y0 , z0 , 1), (v1 , v2 , v3 , 0)i, we have:
Definition 6.7.7. A bundle of lines in A3 (K) is the set of all lines that
pass through a given point.
An improper bundle of lines in A3 (K) is the set of all lines that are
parallel to a given one.
A bundle of planes in A3 (K) is the set of all planes that pass through
a given point.
An improper bundle of planes in A3 (K) is the set of all planes that
are parallel to a given line.
153
Andrea Ferraguti Chapter 6: Affine geometry
where (λ, µ, ν) ∈ K 3 \ {(0, 0, 0)}. Notice that two non-zero triples (λ0 , µ0 , ν0 )
and (λ00 , µ00 , ν00 ) in K 3 determine the same line in the bundle if and only if they
are proportional. Equivalently, one can write down a cartesian equation for
the bundle of lines through P , that is the following:
(
µ(x − xP ) − λ(y − yP ) = 0
.
ν(x − xP ) − λ(z − zP ) = 0
Given a line ` ⊆ A3 (K) with translation space generated by a non-zero
vector (v1 , v2 , v3 ) ∈ K 3 , the parametric equation for the bundle of lines parallel
to ` is:
x = λ + v1 t
y = µ + v2 t ,
z = ν + v3 t
where (λ, µ, ν) ∈ K 3 . Two triples (λ0 , µ0 , ν0 ) and (λ00 , µ00 , ν00 ) in K 3 determine
the same line in the bundle if and only if the line through the points (λ0 , µ0 , ν0 )
and (λ00 , µ00 , ν00 ) of A3 (K) has direction (v1 , v2 , v3 ). This happens precisely when
λ0 − λ00 µ0 − µ00 ν0 − ν00
rk = 1.
v1 v2 v3
The proof of the following proposition is completely analogous to that of
Proposition 6.7.2, so we omit it. We invite the interested readers to try to
write it down themselves.
Proposition 6.7.8.
154
Andrea Ferraguti Chapter 6: Affine geometry
155
Andrea Ferraguti Chapter 7: Euclidean geometry
Example 7.1.2. The affine space An (R) can be seen as an euclidean space
of dimension n when the underlying vector space Rn is endowed with the
standard scalar product. This euclidean space is denoted by En (R).
Note:-
In this chapter we will appeal several times, without citing it every time,
to the following facts. Let V be an R-vector space of dimension n with a
positive definite scalar product • and W ⊆ V be a vector subspace. Then:
• W ⊕ W⊥ = V ;
• dim W ⊥ = n − dim W ;
• (W ⊥ )⊥ = W .
These facts are the content of Theorem 4.2.13 and Corollary 4.2.14.
156
Andrea Ferraguti Chapter 7: Euclidean geometry
157
Andrea Ferraguti Chapter 7: Euclidean geometry
Proposition 7.1.8. Let `, `0 ⊆ E3 (R) be two skew lines. Then there exists
a unique line that is orthogonal and incident to both ` and `0 .
w • (w0 − w) = w0 • (w0 − w) = 0,
and subtracting the two equations term by term and using the properties
of the scalar product we would get
kw0 − wk2 = 0,
158
Andrea Ferraguti Chapter 7: Euclidean geometry
P Q = {tλ−
−→ (P ) : λ ∈ [0, 1]},
PQ
159
Andrea Ferraguti Chapter 7: Euclidean geometry
coordinate system is chosen then Theorems 6.4.2 and 6.4.5 can be applied,
and E can be ”transformed” into the well-known affine space An (R). However,
this time we have an additional piece of structure, that is a positive definite
scalar product. Hence it is natural to ask what happens to the scalar product
on V , once we apply the map Λ[O,B] .
ΦB : V → Rn
v 7→ (α1 , . . . , αn )
Pn
where v = i=1 αi v i be the bijection of Theorem 6.4.2. Then for every
v, w ∈ V we have:
v • w = ΦB (v) · ΦB (w),
where the scalar product on the right hand side is the standard scalar
product on Rn .
where in the second equality we just used the properties of scalar products
and in the third equality we used the fact that B is an orthonormal basis.
160
Andrea Ferraguti Chapter 7: Euclidean geometry
Proposition 7.2.4.
161
Andrea Ferraguti Chapter 7: Euclidean geometry
a1 (xP1 + a1 t) + . . . + an (xPn + an t) + a0 = 0,
namely t = − PH(P
n
)
a2
, where we set H(P ) = a1 xP1 + a2 xP2 + . . . + an xPn + a0 .
i=1 i
162
Andrea Ferraguti Chapter 7: Euclidean geometry
−→
Hence the vector P Q is just
−→ H(P ) H(P ) H(P )
P Q = − a1 Pn 2 , a2 Pn 2 , . . . , an Pn 2
i=1 ai i=1 ai i=1 ai
Definition 7.2.8.
163
Andrea Ferraguti Chapter 7: Euclidean geometry
d(`, `0 ) = d(π, π 0 ).
Proof. Since π k π 0 , the two planes have the same translation space W .
Since ` ⊆ π and `0 ⊆ π 0 , if U is the translation space of ` and U 0 is the
translation space of `0 then necessarily U + U 0 ⊆ W . On the other hand
since ` and `0 are not parallel then U 6= U 0 and hence U ⊕ U 0 = W . In
particular, if (u) is a basis of U and (u0 ) is a basis of U 0 , then (u, u0 ) is a
basis of W .
Let r be the unique line that is incident and orthogonal to both ` and
`0 , and let Ur = hur i be its translation space. Since r ⊥ ` and r ⊥ `0 , we
have that ur • u = 0 and ur • u0 = 0; it follows that ur ∈ W ⊥ . Namely,
r ⊥ π. Let P = ` ∩ r and Q = `0 ∩ r. Then Q is the orthogonal projection
of P on π 0 , because it is the intersection of the unique line orthogonal to
π 0 and passing through P , that is r. Hence d(P, Q) is both the distance
from ` to `0 and that of π from π 0 .
Example 7.2.12. Let us compute the distance between the two lines
( (
x+y =0 0 x−y =0
`: and ` : .
x−z =1 2x + z = 0
By solving the two systems one sees that the translation space W of ` is
generated by (1, −1, 1) while the translation space W 0 of `0 is generated
by (1, 1, −2). Hence
⊥
W ⊥ = h(1, 1, 0), (1, 0, −1)i and W 0 = h(1, −1, 0), (2, 0, 1)i.
164
Andrea Ferraguti Chapter 7: Euclidean geometry
and its intersection with ` is the point P = (1/7, −1/7, −6/7). Hence we
have
q
d(`, `0 ) = d(P, Q) = (2/7 − 1/7)2 + (2/7 + 1/7)2 + (−4/7 + 6/7)2 =
√ √
14 2
= =√ .
7 7
Now let us try to compute d(`, `0 ) using Proposition 7.2.11. First, we
need to find two parallel planes π, π 0 such that ` ⊆ π and `0 ⊆ π 0 . To do
this, notice that every plane of the form a(x + y) + b(x − z − 1) = 0, for
a, b ∈ R, contains `, and every plane of the form c(x − y) + d(2x + z) = 0,
for c, d ∈ R, contains `0 . Rewrite these equations as:
165
Andrea Ferraguti Chapter 7: Euclidean geometry
166
Andrea Ferraguti Chapter 8: Projective geometry
Example 8.1.2.
167
Andrea Ferraguti Chapter 8: Projective geometry
[s] := {t ∈ S : s ∼ t}
Proof. Suppose that [s] ∩ [t] 6= ∅, so that there exists u ∈ [s] ∩ [t]. Now
let v ∈ [s]. Then by definition v ∼ s. On the other hand u ∈ [s] as well,
so that u ∼ s. Since the relation R is symmetric and transitive, it follows
that v ∼ u. On the other hand u ∈ [t], and hence u ∼ t. Since the relation
R is transitive, it follows that v ∼ t, and hence v ∈ [t]. This shows that
[s] ⊆ [t]. A completely symmetric argument shows that [t] ⊆ [s], and
hence [s] = [t].
Proposition 8.1.5 shows that an equivalence relation on S defines a partition
on S. That is, we can ”slice” S into equivalence classes that have empty
intersection, and are such that every element of S belongs to exactly one
equivalence class.
Example 8.1.7.
168
Andrea Ferraguti Chapter 8: Projective geometry
169
Andrea Ferraguti Chapter 8: Projective geometry
Let us try to understand in detail elements of the projective space. The key
observation is the following: suppose that (x1 , . . . , xn+1 ) ∈ K n+1 \ {0} is such
that xn+1 6= 0. Then
That is, whenever the last entry of (x1 , . . . , xn+1 ) is non-zero, the equivalence
class of (x1 , . . . , xn+1 ) contains an element whose last coordinate is 1. On the
other hand, suppose that (x1 , . . . , xn , 1), (y1 , . . . , yn , 1) ∈ K n+1 \ {0}. Then
these two elements are either equal or they are not in relation with each other.
In fact, if they were then there would be some non-zero λ ∈ K such that
but λ(x1 , . . . , xn , 1) = (λx1 , . . . , λxn , λ), and if this equals (y1 , . . . , yn , 1) then
necessarily λ = 1. But then (x1 , . . . , xn , 1) = (y1 , . . . , yn , 1).
Hence we can easily prove the following proposition.
φ : K n → Pn (K)
φ((x−1 −1 −1 −1
n+1 x1 , . . . , xn+1 xn )) = [(xn+1 x1 , . . . , xn+1 xn , 1)] = [X],
170
Andrea Ferraguti Chapter 8: Projective geometry
Proposition 8.2.4. There exists a bijection between Pn−1 (K) and the set
of equivalence classes of elements (x1 , . . . , xn+1 ) ∈ K n+1 \ {0} with xn+1 =
0.
From now on, we will denote equivalence classes of the form [(x1 , . . . , xn+1 )]
by (x1 : x2 : . . . : xn+1 ).
We have seen that the space Pn (K) decomposes into two parts: the proper
points and the improper points. Let us now understand them more in detail
in the cases n = 1, 2, 3.
For n = 1, the proper points are in bijection with K, and they are in
bijection with points of the form (x : 1), where x is an element of K. What
are improper points? They are points of the form (x : 0), where x is a non-
zero element of K. But of course any two elements (x, 0), (y, 0) ∈ K 2 \ {0}
are in relation with each other, since x−1 y(x, 0) = (y, 0). Therefore there is
a unique improper point, that is (1 : 0). Hence P1 (K) is the union of K and
an improper point. This should be thought as a ”point at infinity”, and it is
sometimes denoted by ∞. Hence we have, as sets, P1 (K) = K ∪ {∞}.
On the other hand, K can be thought as the set of points of the affine
space A1 (K), and hence we can write
171
Andrea Ferraguti Chapter 8: Projective geometry
Remark 8.2.6. Another way we can think about points in P1 (K), without
distinguishing between proper and improper points, is as 1-dimensional
vector subspaces of K 2 . In fact, every non-zero (x, y) ∈ K 2 generates
a 1-dimensional subspace, that coincides with the subspace generated by
λ(x, y), for every λ ∈ K \ 0. Conversely, every 1-dimensional subspace is
generated by a non-zero vector of K 2 . Therefore there is a bijection
The projective space P2 (K) is called projective plane. Its proper points are
in bijection with K 2 , and in turn they are in bijection with points of the form
(x : y : 1), where (x, y) ∈ K 2 . Since K 2 is the set of points of A2 (K), we can
say that proper points are in bijection with A2 (K). Improper points are points
of the form (x : y : 0), and they are in bijection with P1 (K). Thanks to Remark
8.2.6, we can think of these as 1-dimensional subspaces of K 2 . Now every line
in A2 (K) has a translation space, that is simply a 1-dimensional subspace of
K 2 . Hence improper points of P2 (K) can be thought as translation spaces of
the lines in A2 (K). In other words there is a bijection
Hence the projective plane should be thought as the affine plane A2 (K) to-
gether with some ”extra points” that represent directions of the lines in A2 (K).
This view extends to higher dimensions as well, since K n \ {0}/ ∼ can
always be thought as the set of 1-dimensional subspaces of K n . Hence for
example there is a bijection
172
Andrea Ferraguti Chapter 8: Projective geometry
Example 8.3.2.
173
Andrea Ferraguti Chapter 8: Projective geometry
or, in other words, what are the proper points of ` and what are the improper
points. Let us start by assuming that (a, b) 6= (0, 0). To find proper points, we
just have to find which points (x0 : y0 : 1) of P2 (K) satisfy ax0 + bx0 + c = 0.
These are clearly in bijection with points on the affine line ax + by + c = 0.
On the other hand, improper points of ` are points of the form (x0 : y0 : 0)
that satisfy
ax0 + by0 = 0
It is immediate to see that there is just one such point, and it is (−b : a : 0).
Moreover, the space h(−b, a)i is precisely the translation space of the line
ax + by + c = 0. All in all, we have proven that the line ax + by + cz has
two types of points: its proper points are essentially points of the affine line
ax + by + c = 0, its unique improper point is (−b : a : 0), and it represents the
direction of the line ax + by + c = 0 in A2 (K).
When (a, b) = (0, 0) the line ` becomes z = 0.
The improper line contains no proper points but it contains all improper
points of P2 (K).
An analogous reasoning can be applied to planes in P3 (K). If π : ax + by +
cz + dt = 0 is a plane with (a, b, c) 6= (0, 0, 0), its proper points are essentially
the points on the affine plane ax + by + cz + d = 0, while its improper points
are the points of the form (x0 : y0 : z0 : 0) that satisfy ax0 + bx0 + cz0 = 0.
Of course vectors (x ∈ K 3 that satisfy such relation constitute the
0 , y0 , z0 )
kernel of the matrix a b c , and this is the translation space of the plane
ax + by + cz = 0 in A3 (K). Therefore improper points of π correspond to
directions of the lines that are contained in the plane ax + by + cz = 0.
When (a, b, c) = (0, 0, 0), the plane π becomes t = 0.
The improper plane contains no proper points but it contains all improper
points of P3 (K).
The above discussion can also be reversed. Namely, given a linear subspace
S of An (K), we can find a linear subspace S 0 of Pn (K) such that S 0 ∩ An (K) =
S. This is very easy to do: all we need to do is to homogeneize the equation of
the subspace. A subspace of An (K) is given by a system of linear equations,
in which all variables x1 , . . . , xn appear to the first degree. If we think of these
174
Andrea Ferraguti Chapter 8: Projective geometry
as of the proper points of a linear subspace of Pn (K), this means that each
variable xi has to be thought as xi /xn+1 . For instance, if ax + by + c = 0 is
the equation of a line in A2 (K), then all we need to do is to replace x with
x/z and y with y/z. The equation then becomes ax/z + by/z + c = 0, and
multiplying by z we get ax + by + cz = 0, that is the equation of a line in
P2 (K).
here one should think of λ and µ as varying parameters that cannot be both
zero. In order to obtain a cartesian equation for ` one just has to impose that
x y z
det zP yP zP = 0.
xQ yQ zQ
175
Andrea Ferraguti Chapter 8: Projective geometry
Example 8.4.1.
• Let
x = λ − µ
y = 2λ
z = −λ + 2µ
• Let
x = 2λ + µ
y = 2λ − µ
z=µ
t=λ
176
Andrea Ferraguti Chapter 8: Projective geometry
177
Andrea Ferraguti Chapter 8: Projective geometry
178
Andrea Ferraguti Chapter 8: Projective geometry
Example 8.5.2.
179
Andrea Ferraguti Chapter 8: Projective geometry
180
Andrea Ferraguti Chapter 8: Projective geometry
Definition 8.6.1.
1. A pencil of lines in P2 (K) is the set of all lines through a given point
of P2 (K).
Proposition 8.6.2.
181
Andrea Ferraguti Chapter 8: Projective geometry
Example 8.6.3.
• Consider the pencil of lines through the point (1, 1) ∈ A2 (K). This
is given by the equation:
λ(x − 1) + µ(y − 1) = 0,
λ(x − y − 1) + µ(z − 2) = 0.
Homogeneizing, we get
What happens if, on the other hand, we start with an improper pencil of
lines/planes? Suppose first we have an improper pencil of lines in A2 (K), that
is given by the equation:
ax + by + µ = 0,
where (a, b) ∈ K 2 \ {(0, 0)} are given coefficients and µ ∈ K is a varying
parameter. Homogeneizing this, we get:
ax + by + µz = 0. (37)
Now this closely recalls the equation of a pencil of lines in P2 (K), except for
the fact that we have only one parameter µ. However, consider the equation
λ(ax + by) + µz = 0, (38)
where λ, µ ∈ K are not both zero. This is precisely the equation of a pencil
of lines in P2 (K). The lines z = 0 and ax + by = 0 intersect in the point
182
Andrea Ferraguti Chapter 8: Projective geometry
(−b : a : 0), so (38) is an equation for the pencil of lines through (−b : a : 0).
Notice that this is the improper point corresponding to the direction of the
affine line ax + by + µ = 0!
How different is this from (37)? Not much: if λ 6= 0, we can divide (38)
by λ, obtaining a line that belongs to the family (37). If λ = 0, on the other
hand, we obtain the line z = 0, that does not appear in (37) for any value of
µ. Hence (38) only differs from (37) by the improper line. We have therefore
understood what improper pencils represent in the projective space: they are
simply pencils of lines through an improper point, and thus the name.
Similarly, consider an improper pencil of planes in A3 (K) given by
ax + by + cz + µ = 0,
λ(ax + by + cz) + µt = 0.
This is the pencil defined by the planes ax+by+cz = 0 and t = 0, that intersect
(as already explained in this chapter) in the line ax + by + cz = t = 0, that
lies on the improper plane and contains all points corresponding to directions
of lines lying on ax + by + cz = 0. This explains what an improper pencil of
planes becomes in the projective space: it simply becomes the pencil of planes
that contain the line ax + by + cz = t = 0, plus the improper plane itself.
This is why in the projective space there is no distinction between proper
and improper pencils: they are defined exactly in the same way, but the latter’s
support is contained in the improper line/plane.
183
Andrea Ferraguti Chapter 8: Projective geometry
Of course, the same argument we used for pencils applies to bundles as well:
in the projective space there is no distinction between proper and improper
bundles because an improper bundle of lines/planes is simply a bundle of
lines/planes passing through an improper point.
Lemma 8.7.3. A point P = (x1 : . . . : xn+1 ) ∈ Pn+1 (C) is real if and only
if P = P .
184
Andrea Ferraguti Chapter 8: Projective geometry
P = (x−1 −1 −1 −1 −1
i x1 : xi x2 : . . . : xi xi−1 : 1 : xi xi+1 : . . . : xi xn+1 )
and
P = (x−1 −1 −1 −1 −1
i x1 : xi x2 : . . . : xi xi−1 : 1 : xi xi+1 : . . . : xi xn+1 ).
Since these points are equal, by definition there exists λ ∈ C such that
λ(x−1 −1 −1 −1 −1
i x1 , xi x2 , . . . , xi xi−1 , 1, xi xi+1 , . . . , xi xn+1 ) = P =
= (x−1 −1 −1 −1 −1
i x1 , xi x2 , . . . , xi xi−1 , 1, xi xi+1 , . . . , xi xn+1 ).
The i-th entry of the left hand side is λ, while the i-th entry of the right
hand side is 1. Therefore λ = 1, and x−1 −1
i xj = xi xj for every j. This
−1
means that xi xj ∈ R for every j ∈ {1, . . . , n}, and hence P is real.
P ∈ L ⇐⇒ P ∈ L.
185
Andrea Ferraguti Chapter 8: Projective geometry
x
1
Pn (K). Then P ∈ L if and only if A . . . = 0, i.e. if and only if
xn+1
(x1 , . . . , xn+1 ) ∈ ker A. Since L is real then ker A = ker A, so this last
condition is equivalent to (x1 , . . . , xn+1 ) ∈ ker A. Taking conjugates, this
becomes equivalent to P ∈ L.
Conversely, suppose that P ∈ L if and only if P ∈ L. Suppose that
(x1 , . . . , xn+1 ) ∈ ker A. Then (x1 : . . . : xn+1 ) ∈ L, and hence (x1 :
. . . : xn+1 ) ∈ L. This implies that (x1 , . . . , xn+1 ) ∈ ker A, and taking
conjugates (x1 : . . . : xn+1 ) ∈ ker A. Therefore ker A ⊆ ker A. The
symmetric argument shows that ker A ⊆ ker A.
186
Andrea Ferraguti Chapter 8: Projective geometry
A= 1 b0 c0 and hence
−b0
A 1 = −b0 + b0 ,
0
which is not 0 since b0 is not real. This contradicts the fact that ker A =
ker A, and hence (a : b : c) must be real.
2. Let P = (x1 : x2 : x3 ) ∈ P2 (C) be an imaginary point, and let P be
its conjugate. Let ` : ax + by + cz = 0 be a line through P and P . Then
(
ax1 + bx2 + cx3 = 0
.
ax1 + bx2 + cx3 = 0
187
Andrea Ferraguti Chapter 8: Projective geometry
The proof of the next proposition is very similar to that of Proposition 8.7.6.
We omit it, but encourage the interested reader in writing it down them-
selves.
7. An imaginary line contains at most one real point and there exists
at most one real plane containing it.
188
Andrea Ferraguti Chapter 9: Conics
Chapter 9: Conics
9.1 Algebraic curves, intersection multiplicities and tan-
gents
We denote by C[x, y, z] the set of polynomials in 3 variables over the complex
field. That is, elements of C[x, y, z] are expressions of the form
n
X
aijk xi y j z k .
i,j,k=1
Example 9.1.2.
Since all of our curves will be contained in P2 (C), we will drop the adjective
”plane”.
189
Andrea Ferraguti Chapter 9: Conics
Note:-
The definition of algebraic curve makes sense only because f is homoge-
neous. In fact, if (x1 : x2P: x3 ) is such that f (x1 , x2 , x3 ) = 0, with f
i j k
homogeneous of the form i+j+k=d aijk x y z , then for every λ ∈ C we
have X
f (λx1 , λx2 , λx3 ) = aijk (λx1 )i (λx2 )j (λx3 )k =
i+j+k=d
X X
= λi+j+k aijk xi1 xj2 xk3 = λd aijk xi1 xj2 xk3 = λd f (x1 , x2 , x3 ) = 0.
i+j+k=d i+j+k=d
Example 9.1.6.
x2 + xy + xz + yz = (x + y)(x + z).
190
Andrea Ferraguti Chapter 9: Conics
The equation
f (x, y, αx + βy) = 0
certainly has a solution, because f (x0 , y0 , αx0 + βy0 ) = f (x0 , y0 , z0 ) = 0. On
the other hand, we have that
X
f (x, y, αx + βy) = aijk xi y j (αx + βy)k .
i+j+k=0
mial of degree k. Hence every term aijk xi y j (αx + βy)k of f (x, y, αx + βy)
is homogeneous of degree d, and in turn the polynomial f (x, y, αx + βy) is
homogeneous of degree d as well. Hence we can write
d
X
f (x, y, αx + βy) = ah xh y d−h (39)
h=0
191
Andrea Ferraguti Chapter 9: Conics
Note:-
Since Equation (40) has degree d, the intersection multiplicity is at most d.
Moreover, it is also at least 1.
The definitoin of intersection multiplicity looks intricate at first, but in fact
computing it is rather easy. Let us see some examples.
Example 9.1.9.
iy + y 2 = 0.
x2 + y 2 − x2 = y 2 = 0. (42)
192
Andrea Ferraguti Chapter 9: Conics
y 2 x = 0.
Referring to 9.1.9, in the first example the line is not tangent, while in the
other two it is.
Example 9.1.12.
x3 − x2 z − m2 x2 z = 0.
x3 − x2 − m2 x2 = x3 − (1 + m2 )x2 = 0.
x3 − xz 2 − m2 x2 z = 0,
193
Andrea Ferraguti Chapter 9: Conics
y 2 z = 0,
194
Andrea Ferraguti Chapter 9: Conics
f (x, y, αx + βy) = 0.
Therefore the theorem is proved if we can show that equation (44) has
exactly d solutions, when counted with multiplicity.
Suppose that (x0 , y0 ) is a solution of (44). We can assume that (x0 , y0 ) 6=
(0, 0) because if x0 = 0 = y0 then z = αx0 + βy0 = 0, which does not
define a point of P2 (C). Now suppose that y0 6= 0. Then (x0 y0−1 , 1) is also
a solution of (44). On the other hand, if (x0 , 0) is a solution than also
(1, 0) is. Since proportional pairs (x0 , y0 ) and (λx0 , λy0 ) give rise to the
same point P = (x0 : y0 : αx0 + βy0 ),P we can just count solution of (44)
of the form (x0 , 1) or (1, 0). Now, if dh=0 ah xh y d−h = ad y d then we are
done, since (1, 0) is the unique root and it has multiplicity d. Otherwise,
195
Andrea Ferraguti Chapter 9: Conics
with ad−e 6= 0. Then (1, 0) is a root with multiplicity e, and the equation
d
X
ah xh y d−h−e = 0
h=0
has precisely d−e roots of the form (x0 , 1), when counted with multiplicity.
All in all, we have e + d − e = d roots.
9.2 Conics
Definition 9.2.1. A conic is a plane algebraic curve of degree 2.
196
Andrea Ferraguti Chapter 9: Conics
Proof. Let
197
Andrea Ferraguti Chapter 9: Conics
198
Andrea Ferraguti Chapter 9: Conics
But both factors have x−1 0 y0 as a root, and hence mP (C, `) ≥ 2, contra-
dicting the fact that C is generic.
2. We have proved in 1. that if C decomposes as the product of two
lines, the intersection point is singular. Hence if f (x, y, z) = (ax + by +
cz)(a0 x + b0 y + c0 z) with ax + by + cz = 0 and a0 x + b0 y + c0 z = 0 distinct
lines, their unique intersection point P is singular. On the other hand, no
other point Q can be singular, as otherwise if ` is the line through P, Q
then mp (C, `) + mQ (C, `) ≥ 3, contradicting Theorem 9.1.15. Conversely,
suppose that C is simply degenerate. Then by 1. it must be reducible,
and since deg f = 2, we can only have f (x, y, z) = g(x, y, z)h(x, y, z) with
g, h homogeneous polynomials of degree 1. If g = 0 and h = 0 define the
199
Andrea Ferraguti Chapter 9: Conics
same line, then any point on such line is singular, but then every point of
C would be singular. Hence g = 0 and h = 0 must define distinct lines.
3. If f = (ax + by + cz)2 , then every point of C is singular, and so
C is doubly degenerate. Conversely, if C is doubly degenerate then it is
reducible, and we showed in 2. that if it decomposes as the product of
two distinct lines then there is a unique singular point. Then it must
decompose as the square of a line.
The next lemma shows how to compute tangent lines in smooth points of
conics.
a a a13
11 12
Lemma 9.2.5. Let C : t XAX = 0 be a conic, where A = a12 a22 a23 ∈
a13 a23 a33
M3 (C). Let P = (xP : yP : zP ) ∈ C be a smooth point. Then the unique
tangent in P is the line
x
xP yP zP A y = 0.
z
200
Andrea Ferraguti Chapter 9: Conics
201
Andrea Ferraguti Chapter 9: Conics
Example 9.2.8.
that is, √
i 2x + 2y = 0.
Similarly, the line √
−i 2x + 2y = 0
is the tangent in P . Since these lines are proper, they are the
asymptotes of C.
202
Andrea Ferraguti Chapter 9: Conics
By Remark 9.2.7, ellipses and hyperbolas have two asymptotes, while parabo-
las have no asymptotes.
Let
a a
e = 11 12 .
A
a12 a22
Then C is:
1. an ellipse if det A
e > 0;
2. a hyperbola if det A
e < 0;
3. a parabola if det A
e = 0.
that is,
a11 x2 + 2a12 xy + a22 y 2 = 0. (48)
Now we need to consider three cases. First, if a11 = a22 = 0 then it
203
Andrea Ferraguti Chapter 9: Conics
must be a12 6= 0, because if it was also a12 = 0 then the first two rows
of A would be linearly dependent, so that rk(A) < 3 and A would not
be general. Then the equation becomes xy = 0, so the conic has two
improper points (1 : 0 : 0) and (0 : 1 : 0), hence it is a hyperbola and
e = −a2 < 0, as required.
det A 12
Next, if a11 6= 0 then y = 0 does not yield a solution of (48). Hence
we can assume that y = 1 and look at the equation
that has two real solutions precisely when ∆ = a212 − a11 a22 > 0, it has one
solution with multiplicity two when a212 −a11 a22 = 0 and has two imaginary
conjugate solutions when a212 − a11 a22 < 0. Since a212 − a11 a22 = − det A,
e
we are done.
Finally, if a22 6= 0 then x = 0 does not yield any solution of (48), so
we can just set x = 1 and reason like in the previous case.
204
Andrea Ferraguti Chapter 9: Conics
Definition 9.3.5. Let C be a real general conic and let P ∈ P2 (C). The
set of points of P2 (C) that are conjugate to P with respect to C is called
the polar of P with respect to C.
Proof. 1. The polar is the set of points that solve the equation
a11 a12 a13 x
xP yP zP a12 a22 a23 y = 0,
a13 a23 a33 z
205
Andrea Ferraguti Chapter 9: Conics
A(P − λQ) = 0.
206
Andrea Ferraguti Chapter 9: Conics
Proposition 9.3.10. Let C be a general real conic, let P ∈ P2 (C) and let
p be the polar of P with respect to C.
2. If P ∈
/ C, there exist exactly two lines `1 and `2 through P that are
tangent to C, and p is the line through the two points `1 ∩ C and
`2 ∩ C.
207
Andrea Ferraguti Chapter 9: Conics
Proposition
9.3.13.Let C be a general real conic with defining matrix
a a a
11 12 13
A = a12 a22 a23 .
a13 a23 a33
208
Andrea Ferraguti Chapter 9: Conics
(notice that the solution is unique because since C is general, the rows of
A are linearly independent, hence all solutions of (49) are proportional to
each other).
2. System (49) has a solution of the form (xP : yP : 0) if and only if
(xP , yP ) solves the system
(
a11 x + a12 y = 0
. (50)
a12 x + a22 y = 0
209
Andrea Ferraguti Chapter 9: Conics
the two equations are linearly dependent. This means that the solution
is (−a12 : a11 : 0) (notice that it cannot be a11 = a12 = 0 as otherwise we
would also have a22 = 0 and C would be degenerate).
On the other hand, to find the improper points of C we need to solve
210
Andrea Ferraguti Chapter 9: Conics
Proof. Every diameter passes through the center, by Remark 9.3.12. Con-
versely, let ` be a line through the center P of C. By Theorem 9.3.9, the
pole of ` must lie on the polar of P . But this is the improper line, by
definition of center. Hence ` is the polar of an improper point.
Remark 9.3.17. When looking at the affine part of a general real conic C, if
C is not a parabola then its center is proper, and therefore the diameters
will be the pencil of affine lines through such proper point. If C is a
parabola, since the center is improper the affine part of the diameters will
constitute a pencil of parallel lines, all with direction (−a12 : a11 : 0) by
Proposition 9.3.13.
211
Andrea Ferraguti Chapter 9: Conics
of the form:
where aij ∈ R for every i, j and a11 , a12 , a22 are not all 0.
Of course by homogeneizing the equation of a conic in E2 (R) one obtains
the equation of a real conic Ce in P2 (C). However, we maintain the concepts
distinct, in a formal way, in order to be able to talk about orthogonality, that
is a notion that makes no sense in P2 .
Every notion we have seen in the previous sections for conics in P2 (C)
carries over to conics in E2 (R); to make sense of such notions we’ll think of
them as associated to the conic C. e For example, if x2 + y 2 + 1 is a conic in
E2 (R), we can talk of its improper points: these will be the improper points
in P2 (C) of the conic x2 + y 2 + z 2 = 0.
In the Euclidean setting we distinguish (for reasons that will not be treated
in these notes) circles from ellipses, although the former are a special case of
the latter.
with a11 6= 0.
Therefore, in the Euclidean setting we will use the word ”ellipse” to denote
an ellipse that is not a circle.
212
Andrea Ferraguti Chapter 9: Conics
Since C is a hyperbola, we know that this equation will yield two real
distinct improper points: (x0 : y0 : 0) and (x00 : y00 : 0). Let P be the
center of C, that is proper by Proposition 9.3.13. Since the asymptotes are
the lines through P and the improper points, again by Proposition 9.3.13,
their directions are (x0 , y0 ) and (x00 , y00 ). Hence they are orthogonal if and
only if
x0 x00 + y0 y00 = 0. (52)
6 0, then x0 =
If, without loss of generality, we assume that a22 = 6 0 as
otherwise to solve (51) we would also need y0 = 0. For the same reason,
x00 6= 0. Hence (52) can be rewritten as:
y0 y00
· = −1.
x0 x00
y0 y00
Now since x0
and x00
are the roots of the equation
213
Andrea Ferraguti Chapter 9: Conics
1. If C is a circle, all diameters are axes and all proper points of C are
vertices.
The direction of such polar is (xP a12 + yP a22 , −xP a11 − yP a12 ), while that
of the pole is (xP , yP ). Hence in order for them to be orthogonal we need
(xP a12 + yP a22 )xP − (xP a11 + yP a12 )yP = 0 or, in other words,
214
Andrea Ferraguti Chapter 9: Conics
P1 6= P2 .
Let `i be the polar of Pi , for i = 1, 2. If `i was the improper line, then
Pi ∈ `i , which implies that Pi ∈ C by Proposition 9.3.10. But the polar of
a point of C is its tangent, by the same proposition, so that the improper
line would be tangent in an improper point of C. This would imply that
C is a parabola by Remark 9.2.7, contradicting the hypothesis. Therefore,
`1 and `2 are axes of C, and they are distinct since they are the polar of
two distinct points, by Proposition 9.3.6.
Now we need to prove that `i ∩ C consists of two distinct points. Let
i ∈ {1, 2}. If `i was tangent to C in a point Q, then `i would be the
polar of Q by Proposition 9.3.10. But since the pole of `i is improper,
then Q would be improper and its direction would be the same as that
of `i , which is impossible. Hence `i ∩ C consists of two distinct points for
i = 1, 2. If `1 ∩ C contains an improper point P 0 , then by Theorem 9.3.9
the polar of P 0 would pass by P2 . But the polar of P 0 is the tangent line in
P 0 , and since P 0 and P2 are both improper, this means that the improper
line z = 0 is tangent to C. Hence C should a parabola, contradicting the
hypothesis. This shows that `1 ∩ C consists of two distinct proper points,
and with a simmetric argument also `2 ∩ C does.
Let then `1 ∩ C = {Q1 , Q2 } and `2 ∩ C = {R1 , R2 } with Q1 , Q2 , R1 , R2
proper points with Q1 6= Q2 and R1 6= R2 . If it was Q1 = R1 , then
Q1 would be the center of the conic, since all diameters pass through
the center. But Q1 ∈ C, and this contradicts Remark 9.3.14. Hence
{Q1 , Q2 } ∩ {R1 , R2 } = ∅, proving that C has 4 vertices.
3. By Theorem 9.3.13, diameters of a parabola are exactly the lines that
pass through its unique improper point. If a11 6= 0 or a12 6= 0, the improper
point is (a12 : −a11 : 0). Otherwise, we must have a22 6= 0 as otherwise Ce
would be degenerate, and the improper point is (a12 : −a22 : 0). Without
loss of generality, let us assume that we are in the first case. Then all
diameters have equation
a11 x + a12 y + kz = 0,
215
Andrea Ferraguti Chapter 9: Conics
9.3.9, the polar of the vertex, that is tangent to C therein, passes through
(a12 : −a11 : 0), and hence its direction is orthogonal to that of the
axis.
Example 9.4.9.
xy = 0.
x − z = 0 and 2y − z = 0
x − 1 = 0 and 2y − 1 = 0.
216
Andrea Ferraguti Chapter 9: Conics
−x2 − 2xy + y 2 = 0,
√
which yields the two points (1 : 1 ± 2 : 0). Now the axes of C are
the polars of these points, namely the lines:
√ √ √ √
2x + (2 + 2)y = 0 and 2x − (2 − 2)y = 0.
and ( √ √
(( 2 − 1)y)2 − y 2 + 2y( 2 − 1)y + 3 = 0
√ ,
x = ( 2 − 1)y
finding the four points
√ √ √
q q
( 2 + 1) 3 − 3 2 : − 3 − 3 2 : 2 ,
√ √ √
q q
−( 2 + 1) 3 − 3 2 : 3 − 3 2 : 2 ,
√ √ √
q q
( 2 − 1) 3 + 3 2 : 3 + 3 2 : 2 ,
√ √ √
q q
( 2 − 1) 3 + 3 2 : 3 + 3 2 : −2 .
217
Andrea Ferraguti Chapter 9: Conics
x2 − y 2 = 0,
x + y = 0 and z = 0,
218
Andrea Ferraguti Chapter 9: Conics
Notes
219