Culegere Peter
Culegere Peter
Cluj-Napoca
2014
Contents
Introduction iii
1 Matrices 1
1.1 Basic definitions, operations and properties. . . . . . . . . . . . . . . 1
1.2 Determinants and systems of linear equations . . . . . . . . . . . . . 13
1.3 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 23
2 Vector Spaces 28
2.1 Definition of Vector Space and basic properties . . . . . . . . . . . . . 28
2.2 Subspaces of a vector space . . . . . . . . . . . . . . . . . . . . . . . 29
2.3 Basis. Dimension. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
2.4 Local computations . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44
2.5 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
i
CONTENTS ii
Bibliography 153
Introduction
The aim of this book is to give an introduction to linear algebra and at the same
time to provide some applications that might be useful both in practice and theory.
Hopefully this book will be a real help for graduate level students to understand
the basics of this beautiful mathematical subject called linear algebra. Our scope is
twofold: one hand we give a theoretical introduction to this field, which is more than
exhaustive for the need and understanding capability of graduate students, on the
other hand we present fully solved examples and problems that might be helpful in
preparing to exams and also show the techniques used in the art of problem solving
on this field. At the end of every chapter, this work contains several proposed
problems that can be solved using the theory and solved examples stated previously.
iii
1
Matrices
Hence, the elements of a matrix A are denoted by aij , where aij stands for the
number that appears in the ith row and the j th column of A (this is called the pi, jq
entry of A) and the matrix is represented as A “ paij qi“1,m .
j“1,n
We will denote the set of all m ˆ n matrices with entries in F by Mm,n pFq
respectively, when m “ n by Mn pFq. It is worth mentioning that the elements of
1
Basic definitions, operations and properties. 2
Mn pFq are called square matrices. In what follows, we provide some examples.
Example 1.5. Consider the matrices O “ paij qi“1,m P Mm,n pFq having all entries
j“1,n
the zero element of the field F. Then
¨ ˛
0 0 ... 0
˚ ‹
˚ ‹
˚0 0 . . . 0‹
O“˚ ˚ .. ..
‹
.. ‹
˚. . . . . .‹
˝ ‚
0 0 ... 0
Basic definitions, operations and properties. 3
Example 1.6. Consider the matrices A “ paij qi,j“1,m P Mn pFq given by aij “ 0
whenever i ą j, respectively aij “ 0 whenever i ă j. Then
¨ ˛ ¨ ˛
a a ... a1n a 0 ... 0
˚ 11 12 ‹ ˚ 11 ‹
˚ ‹ ˚ ‹
˚ 0 a22 . . . a2n ‹ ˚ a21 a22 ... 0 ‹
A“˚ ˚ .. ..
‹ ˚
.. ‹ , respectively A “ ˚ .. .. ..
‹
‹
˚ . . ... . ‹ ˚ . . ... . ‹
˝ ‚ ˝ ‚
0 0 ... ann an1 an2 . . . ann
Addition of Matrices.
If A and B are m ˆ n matrices, the sum of A and B is defined to be the m ˆ n
matrix A ` B obtained by adding corresponding entries. Hence, the addition
operation is a function
paij qi“1,m ` pbij qi“1,m “ paij ` bij qi“1,m , @ paij qi“1,m , pbij qi“1,m P Mm,n pFq.
j“1,n j“1,n j“1,n j“1,n j“1,n
C “ A ` B “ pcij qi“1,m
j“1,n
where cij “ aij ` bij for all i P t1, 2, . . . , mu, j P t1, 2, . . . , nu.
Basic definitions, operations and properties. 4
2. pA ` Bq ` C “ A ` pB ` Cq (associative property).
3. A ` B “ B ` A (commutative property).
4. A ` O “ O ` A “ A (additive identity).
Scalar multiplication.
For A P Mm,n pFq and α P F define αA “ pαaij qi“1,m . Hence, the scalar
j“1,n
multiplication operation is a function
Of course that we listed here only the left multiplication of matrices by scalars. By
defining αA “ Aα we obtain the right multiplication of matrices by scalars.
¨ ˛ ¨ ˛
1 ´1 1 ´1 0 2
˚ ‹ ˚ ‹
˚ ‹ ˚ ‹
Example 1.7. If A “ ˚ 0 2 ´1 ‹ and B “ ˚ 1 ´1 1 ‹ , then
˝ ‚ ˝ ‚
´2 2 0 0 ´1 2
¨ ˛ ¨ ˛
3 ´2 0 1 ´2 4
˚ ‹ ˚ ‹
˚ ‹ ˚ ‹
2A ´ B “ ˚ ´1 5 ´3 ‹ and 2A ` B “ ˚ 1 3 ´1 ‹ .
˝ ‚ ˝ ‚
´4 5 ´2 ´4 3 2
Transpose.
The transpose of a matrix A P Mm,n pFq is defined to be a matrix AJ P Mn,m pFq
obtaining by interchanging rows and columns of A. Locally, if A “ paij qi“1,m , then
j“1,n
AJ “ paji q j“1,n .
i“1,m
It is clear that pAJ qJ “ A. A matrix, that has many columns, but only one row, is
called a row matrix. Thus, row matrix A with n columns is an 1 ˆ n matrix, i.e.
A “ pa1 a2 a3 . . . an q.
A matrix, that has many rows, but only one column, is called a column matrix.
Thus, a column matrix A with m rows is an m ˆ 1 matrix, i.e.
¨ ˛
a1
˚ ‹
˚ ‹
˚ a2 ‹
˚
A“˚ . ‹ ‹.
˚ .. ‹
˝ ‚
am
Basic definitions, operations and properties. 6
Obviously, the transpose of a row matrix is a column matrix and viceversa, hence,
in inline text a column matrix A is represented as
A “ pa1 a2 . . . am qJ .
Conjugate Transpose.
Let A P Mm,n pCq. Define the conjugate transpose of A “ paij qi“1,m P Mm,n pCq by
j“1,n
A‹ “ paji q j“1,n , where z denotes the complex conjugate of the number z P C. We
i“1,m
have that pA‹ q‹ “ A and AJ “ A‹ whenever A contains only real entries.
1. pA ` BqJ “ AJ ` B J .
2. pA ` Bq‹ “ A‹ ` B ‹ .
Symmetries.
Let A “ paij q i“1,n P Mn pFq be a square matrix. We recall that
j“1,n
Matrix multiplication.
For a matrix X “ pxij qi“1,m P Mm,n pFq we denote by Xi˚ its ith row, i.e. the row
j“1,n
matrix
Xi‹ “ pxi1 xi2 . . . xin q.
It is obvious that
pX J qi‹ “ pX‹i qJ ,
Basic definitions, operations and properties. 8
respectively
pX J q‹j “ pXj‹ qJ .
We say that the matrices A and B are conformable for multiplication in the order
AB, whenever A has exactly as many columns as B has rows, that is A P Mm,p pFq
and B P Mp,n pFq.
For conformable matrices A “ paij qi“1,m and B “ pbjk q j“1,p the matrix product AB
j“1,p k“1,n
is defined to be the m ˆ n matrix C “ pcik qi“1,m with
k“1,n
ÿ
p
cik “ Ai‹ B‹k “ aij bjk .
j“1
In the case that A and B failed to be conformable, the product AB is not defined.
Remark 1.9. Note, the product is not commutative, that is, in general,
AB ‰ BA even if both products exists and have the same shape.
¨ ˛
¨ ˛ 1 ´1
1 0 ´1 ˚ ‹
˚ ‹
Example 1.10. Let A “ ˝ ‚ and B “ ˚ 0 1 ‹.
´1 1 0 ˝ ‚
´1 1
¨ ˛
¨ ˛ 2 ´1 ´1
2 0 ˚ ‹
Then AB “ ˝ ‚ and BA “ ˚ ˚ ´1 1
‹
0 ‹.
´1 2 ˝ ‚
0 1 1
There are some similar representations for individual columns, i.e. the j th column
is
Consequently, we have:
The last two equations has both theoretical and practical importance. They shows
that the rows of AB are combinations of rows of B, while columns of AB are
combinations of columns of A. So it is waisted time to compute the entire product
when only one row or column is needed.
Basic definitions, operations and properties. 10
AIn “ A and In A “ A ,
Proposition 1.11. For conformable matrices A P Mm,p pFq and B P Mp,n pFq, on
has
pABqJ “ B J AJ .
pABq‹ “ B ‹ A‹ .
Exercise. Prove, that for every matrix A “ paij qi“1,m P Mm,n pFq the matrices
j“1,n
AAJ and AJ A are symmetric matrices.
For a matrix A P Mn pFq, one can introduce its mth power by
A0 “ In , A1 “ A, Am “ Am´1 A.
¨ ˛ ¨ ˛ ¨ ˛
0 1 ´1 0 0 ´1
Example 1.12. If A “ ˝ ‚ then A2 “ ˝ ‚, A3 “ ˝ ‚
´1 0 0 ´1 1 0
¨ ˛
1 0
and A4 “ ˝ ‚ “ I2 . Hence Am “ Amp mod q4 .
0 1
Proposition 1.13. For A P Mm,n pCq and B P Mn,m pCq one has
trace AB “ trace BA.
Proof. We have
ÿ
m ÿ
m ÿ
m ÿ
n
trace AB “ rABsii “ pAqi‹ pBq‹i “ aik bki “
i“1 i“1 i“1 k“1
ÿ
m ÿ
n ÿ
n ÿ
m ÿ
n
bki aik “ bki aik “ rBAskk “ trace BA.
i“1 k“1 k“1 i“1 k“1
¨ ˛ ¨ ˛
A A12 ... A1r B B12 ... B1t
˚ 11 ‹ ˚ 11 ‹
˚ ‹ ˚ ‹
˚A21 A22 ... A2r ‹ ˚B B22 . . . B2r ‹
A“˚ ‹ and B “ ˚ 21 ‹
˚ .. .. .. ‹ ˚ .. .. .. ‹
˚ . . ... . ‹ ˚ . . ... . ‹
˝ ‚ ˝ ‚
As1 As2 ... Asr Br1 Br2 . . . Brt
We say that the partitioned matrices are conformable partitioned if the pairs
pAik , Bkj q are conformable matrices, for every indices i, j, k. In this case the
product AB is formed by combining blocks exactly the same way as the scalars are
combined in ordinary matrix multiplication. That is, the pi, jq block in the
product AB is
Ai1 B1j ` Ai2 B2j ` . . . Air Brj .
Matrix Inversion.
For a square matrix A P Mn pFq, the matrix B P Mn pFq that satisfies
AB “ In and BA “ In
(if exists) is called the inverse of A and is denoted by B “ A´1 . Not all square
matrices admits an inverse (are invertible). An invertible square matrix is called
nonsingular and a square matrix with no inverse is called singular matrix.
Although not all matrices are invertible, when an inverse exists, it is unique.
Indeed, suppose that X1 and X2 are both inverses for a nonsingular matrix A.
Then
1. pA´1 q´1 “ A
3. pABq´1 “ B ´1 A´1 .
It can easily be computed, that for A “ paij q i“1,2 P M2 pFq, one has
j“1,2
Similarly, if A “ paij q i“1,3 P M3 pFq, then its determinant can be calculated by the
j“1,3
rule
detpAq “
a11 a22 a33 ` a13 a21 a32 ` a12 a23 a31 ´ a13 a22 a31 ´ a11 a23 a32 ´ a12 a21 a33 .
¨ ˛
1 2 3
˚ ‹
˚ ‹
Example 1.15. If A “ ˚ 4 5 6 ‹ then
˝ ‚
7 8 9
detpAq “ 1 ¨ 5 ¨ 9 ` 3 ¨ 4 ¨ 8 ` 2 ¨ 6 ¨ 7 ´ 3 ¨ 5 ¨ 7 ´ 1 ¨ 6 ¨ 8 ´ 2 ¨ 4 ¨ 9 “ 0.
Determinants and systems of linear equations 15
Laplace’s theorem.
Let A P Mn pFq and let k be an integer, 1 ď k ď n. Consider the rows i1 . . . ik and
the columns j1 . . . jk of A. By deleting the other rows and columns we obtain a
submatrix of A of order k, whose determinant is called a minor of A and is
denoted by Mij11...i
...jk
k
. Now let us delete the rows i1 . . . ik and the columns j1 . . . jk of
A.. We obtain a submatrix of A of order n ´ k. Its determinant is called the
complementary minor of Mij11...i
...jk Ăj1 ...jk . Finally let us denote
and it is denoted by M
k i1 ,...ik
Aij11...i
...jk Ăj1 ...jk .
“ p´1qi1 `¨¨¨`ik `j1 `¨¨¨`jk M
k i1 ...ik
˜ ¸J
The adjugate of A is the matrix adjpAq “ pAji q i“1,n , that is
j“1,n
¨ ˛
A11 A12 ¨¨¨ A1n
˚ ‹
˚ ‹
˚ A21 A22 ¨¨¨ A2n ‹
adjpAq “ ˚
˚ .. .. .. ‹
‹
˚ . . ¨¨¨ . ‹
˝ ‚
An1 An2 ¨ ¨ ¨ An n
1
A´1 “ adjpAq.
detpAq
Theorem 1.18.
ÿ
detpAq “ ...jk j1 ...jk
Mij11...i k
Ai1 ...ik , where
• The indices j1 . . . jk runs over all the possible values, such that
1 ď j1 ă ¨ ¨ ¨ ă jk ď n.
(3) A determinant with two equal rows (or two equal columns) is zero.
(4) The determinant of A is not changed if a multiple of one row (or column) is
added to another row (or column).
(5) detpA´1 q “ 1
detpAq
.
rank pAq ` rank pBq ´ m ď rank pABq ď mintrank pAq, rank pBqu .
Theorem 1.20. Let A, B P Mn pFq with detpAq ‰ 0. Then rank pABq “ rank pBq.
Proof. Since detpAq ‰ 0, we have rank pAq “ n. By using the above notation with
m “ p “ n we obtain rank pBq ď rank pABq ď rank pBq. Hence
rank pABq “ rank pBq.
• the diagonal entries of B are either 0 or 1, all the 1’s preceding all the 0’s on
the diagonal.
¨ ˛ ¨ ˛
1 2 0 1 0 0
˚ ‹ ˚ ‹ p´A `A q
˚ ‹ ˚ ‹ 2‹ 1‹
˚ 0 2 0 ‹ ˚ 0 1 ´3 ‹ 1 »
˝ ‚ ˝ ‚
0 0 3 0 0 1
¨ ˛ ¨ ˛ ¨ ˛ ¨ ˛
1 0 0 1 ´1 31 1 0 0 1 ´1 1
˚ ‹ ˚ ‹ p 1 A2‹ , 1 A3‹ q ˚ ‹ ˚ 3 ‹
˚ ‹ ˚ 1 ‹ 2 3 ˚ ‹ ˚ ‹
˚ 0 2 0 ‹ ˚ 0 1 ´3 ‹ » ˚ 0 1 0 ‹ ˚ 0 1
´ 61 ‹.
˝ ‚ ˝ ‚ ˝ ‚ ˝ 2 ‚
1
0 0 3 0 0 1 0 0 1 0 0 3
¨ ˛
1
1 ´1 3
˚ ‹
´1 ˚ ‹
Hence A “ ˚ 0 2 ´ 6 ‹ .
1 1
˝ ‚
1
0 0 3
Recall that a matrix is in row echelon form if
(1) All nonzero rows are above any rows of all zeroes.
(2) The first nonzero element (leading coefficient) of a nonzero row is always
strictly to the right of the first nonzero element of the row above it.
2. rank pAq “ n.
4. Ax “ 0 implies that x “ 0.
Here x1 , x2 , . . . , xn are the unknowns, a11 , a12 , . . . , amn are the coefficients of the
system, and b1 , b2 , . . . , bm are the constant terms. Observe that a systems of linear
equations may be written as Ax “ b, with A “ paij qi“1,m P Mm,n pFq, x P Mn,1 pFq
j“1,n
and b P Mm,1 pFq. The matrix A is called the coefficient matrix, while the matrix
rA|bs P Mm,n`1 pFq, $
& a if j ‰ n ` 1
ij
rA|bsij “
% b if j “ n ` 1
i
¨ ˛
1 ´1 0 2 ´2
˚ ‹
˚ ‹
˚ 2 1 ´1 0 4 ‹ p´2A1‹ `A2‹ ,´A1‹ `A3‹ q
We ˚
have rA|bs “ ˚ ‹ »
‹
˚ 1 ´1 ´2 1 1 ‹
˝ ‚
0 1 1 1 ´1
¨ ˛
1 ´1 0 2 ´2
˚ ‹
˚ ‹
˚ 0 3 ´1 ´4 8 ‹ pA2‹ ØA4‹ q
˚ ‹ »
˚ ‹
˚ 0 0 ´2 ´1 3 ‹
˝ ‚
0 1 1 1 ´1
¨ ˛
1 ´1 0 2 ´2
˚ ‹
˚ ‹
˚ 0 1 1 1 ´1 ‹ pA2‹ `A1‹ ,´3A2‹ `A4‹ q
˚ ‹ »
˚ ‹
˚ 0 0 ´2 ´1 3 ‹
˝ ‚
0 3 ´1 ´4 8
Determinants and systems of linear equations 22
¨ ˛
1 0 1 3 ´3
˚ ‹
˚ ‹
˚ 0 1 1 1 ´1 ‹ p 12 A3‹ `A1‹ , 21 A3‹ `A.21‹ ,´2A3‹ `A4‹ q
˚ ‹ »
˚ ‹
˚ 0 0 ´2 ´1 3 ‹
˝ ‚
0 0 ´4 ´7 11
¨ ˛
5 3
1 0 0 ´2
˚ 2 ‹
˚ 1 1 ‹ 1
˚ 0 1 0 ‹ p 2 A4‹ `A1‹ , 101 A4‹ `A2‹ ,´ 15 A4‹ `A3‹ q
˚ 2 2 ‹ »
˚ ‹
˚ 0 0 ´2 ´1 3 ‹
˝ ‚
0 0 0 ´5 5
¨ ˛ ¨ ˛
1 0 0 0 1 1 0 0 0 1
˚ ‹ ˚ ‹
˚ ‹ 1 ˚ ‹
˚ 0 1 0 0 1 ‹ p´ 2 A3‹ ,´ 5 A4‹ q ˚ 0 1 0 0
1
1 ‹
˚ ‹ » ˚ ‹.
˚ ‹ ˚ ‹
˚ 0 0 ´2 0 2 ‹ ˚ 0 0 1 0 ´1 ‹
˝ ‚ ˝ ‚
0 0 0 ´5 5 0 0 0 1 ´1
One can easily read the solution x1 “ 1, x2 “ 1, x3 “ ´1, x4 “ ´1.
Recall that a system of linear equations is called homogeneous if b “ p0 0 ¨ ¨ ¨ 0qJ
that is $
’
’ a11 x1 ` a12 x2 ` ¨ ¨ ¨ a1n xn “ 0
’
’
’
’
& a x ` a x ` ¨¨¨a x “ 0
21 1 22 2 2n n
’
’ ¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨¨
’
’
’
’
% a x ` a x ` ¨ ¨ ¨ a x “ 0.
m1 1 m2 2 mn n
Ax “ O.
1.3 Problems
Problem 1.3.1. By using Laplace’s
› theorem compute
› the following determinants.
› › › ›
› › › 2 1 0 0 0 0 ›
› 1 2 3 4 5 › › ›
› › › ›
› › › 1 2 1 0 0 0 ›
› 2 1 2 3 4 › › ›
› › › ›
› › › 0 1 2 1 0 0 ›
D1 “ › 0 2 1 2 3 › , D2 “ › › ›.
› › ›
› › › 0 0 1 2 1 0 ›
› 0 0 2 1 2 › › ›
› › › ›
› › › 0 0 0 1 2 1 ›
› 0 0 0 2 1 › › ›
› ›
› 0 0 0 0 1 2 ›
› ›
› ›
› 1 1 1 ... 1 ›
› ›
› ›
› 1 ǫ ǫ2 . . . ǫn´1 ›
› ›
› 2pn´1q ›
b) › 1 ǫ2 ǫ 4
... ǫ › , where ǫ “ cos 2π
n
` i sin 2π
n
.
› ›
› .. .. .. .. .. ›
› . . . . . ›
› ›
› 2 ›
› 1 ǫn´1 ǫ2pn´1q . . . ǫpn´1q ›
a) detpAq “ detpAq.
› ›
› ›
› 1 1 1 ... 1 ›
› ›
› ›
› a1 a2 a3 ... an ›
› ›
› ›
a) › a21 a22 a23 ... a2n ›.
› ›
› .. .. .. .. .. ›
› . . . . . ›
› ›
› n´1 n´1 ›
› a1 a2 a3n´1 . . . ann´1 ›
› ›
› ›
› a1 a2 a3 . . . an ›
› ›
› ›
› an a1 a2 . . . an´1 ›
b) ›› . .. .. ..
›
.. › .
› .. . . . . ›
› ›
› ›
› a2 a3 a4 . . . a1 ›
Problem 1.3.6. Compute the rank of the following matrices by using the
Gauss-Jordan elimination method.
¨ ˛ ¨ ˛
0 1 ´2 ´3 ´5 1 2 ´2 3 ´2
˚ ‹ ˚ ‹
˚ ‹ ˚ ‹
˚ 6 ´1 1 2 3 ‹ ˚ 3 ´1 1 ´3 4 ‹
˚
a) ˚ ‹ ˚ ‹.
‹, ˚ ‹
˚´2 4 3 2 1 ‹ ˚´2 1 0 1 ´1‹
˝ ‚ ˝ ‚
´3 0 2 1 2 2 0 0 ´1 0
Problems 25
¨ ˛
1 ´2 3 5 ´3 6
˚ ‹
˚ ‹
˚0 1 2 3 4 7 ‹
˚ ‹
˚ ‹
b) ˚2 1 3 3 ´2 5 ‹ .
˚ ‹
˚ ‹
˚5 0 9 11 ´7 16‹
˝ ‚
2 4 9 12 10 26
Problem 1.3.7. Find the inverses of the following matrices by using the
Gauss-Jordan elimination method.
¨ ˛
¨ ˛ 2 ´1 1
1 1 ˚ ‹
˝ ‚ ˚ ‹
a) A “ , B “ ˚1 2 3 ‹.
1 3 ˝ ‚
3 1 ´1
$
& 1 if i ‰ j
b) A “ paij q i“1,n P Mn pRq, where aij “
j“1,n
% 0 otherwise.
Problem 1.3.8. Prove that if A and B are square matrices of the same size, both
invertible, then:
Problem 1.3.9. For every matrix A P Mm,n pCq prove that the product A‹ A and
AA‹ are hermitian matrices.
Problem 1.3.10. For a quadratic matrix A of order n explain why the equation
AX ´ XA “ I
has no solution.
a) $
’
’ 2x1 ´ 3x2 ` x3 ` 4x4 “ 13
’
’
’
’
& 3x1 ` x2 ´ x3 ` 8x4 “ 2
’
’ 5x1 ` 3x2 ´ 4x3 ` 2x4 “ ´12
’
’
’
’
% x ` 4x ´ 2x ` 2x “ ´12.
1 2 3 4
b) $
’
’ x1 ´ x2 ` x3 ´ x4 ` x5 ´ x6 “ 1
’
’
’
’
’
’ x1 ` x2 ` x3 ` x4 ` x5 ` x6 “ 1
’
’
’
’
& 2x1 ` x3 ´ x5 “ 1
’
’ x2 ´ 3x3 ` 4x4 “ ´4
’
’
’
’
’
’ ´x1 ` 3x2 ` 5x3 ´ x6 “ ´1
’
’
’
’
% x ` 2x ` 3x ` 4x ` 5x ` 6x “ 2
1 2 3 4 5 6
a) $
’
’ 2x ´ y ´ z “ 0
’
’
’
’
’
’ x ` 2y ´ 3z “ 0
’
’
’
’
& 2x ` 3y ` mz “ 0
’
’ nx ` y ` z “ 0
’
’
’
’
’
’ x ` py ` 6z “ 0
’
’
’
’
% 2ex “ y ` z ` 2.
b) $
’
’ 2x ´ y ` z “ 0
’
’
’
’
’
’ ´x ` 2y ` z “ 0
’
’
’
’
& mx ´ y ` 2z “ 0
’
’ x ` ny ´ 2z “ 0
’
’
’
’
’
’ 3x ` y ` pz “ 0
’
’
’
’
% x2 ` y 2 ` x2 “ 3.
2
Vector Spaces
3. αpβvq “ pαβqv
4. 1 ¨ v “ v, @v P V
The elements of V are called vectors and the elements of F are called scalars. The
scalar multiplication depends upon F. For this reason when we need to be exact
28
Subspaces of a vector space 29
we will say that V is a vector space over F, instead of simply saying that V is a
vector space. Usually a vector space over R is called a real vector space and a
vector space over C is called a complex vector space.
Remark. From the definition of a vector space V over F the following rules for
calculus are easily deduced:
• α ¨ 0V “ 0
• 0F ¨ v “ 0V
• α ¨ v “ 0V ñ α “ 0F or v “ 0V .
• V “ FrXs, the set of all polynomials with coefficients in F with the usual
addition and scalar multiplication is an F vector space.
• Mm,n pFq with the usual addition and scalar multiplication is a F vector space.
• Cra,bs , the set of all continuous real valued functions defined on the interval
ra, bs, with the usual addition and scalar multiplication is an R vector space.
• v ´ u P U, @v, u P U
• αv P U, @α P F, @v P U
αv ` βu P U, @α, β P F, @v, u P V.
αv1 ` βv2 P S.
Subspaces of a vector space 31
x1 ` y1 ` z1 “ 0
x2 ` y2 ` z2 “ 0,
and by multiplying the equations with α and β respectively, and adding the
resulting equations we obtain
are subspaces of V .
Proof. We prove the statements by making use of the Proposition 2.4. Let α, β P F
and let u, v P U X W. Then u, v P U and u, v P W. Since U and W are vector spaces
it follows that αv ` βu P U, respectively αv ` βu P W. Hence αv ` βu P U X W.
Now consider α, β P F and let x, y P U ` W. Then x “ u1 ` w1 , y “ u2 ` w2 for
some vectors u1 , u2 P U, w1 , w2 P W. But then
The subspace U X W is called the intersection vector subspace, while the subspace
U ` W is called the sum vector subspace. Of course that these definitions can be
also given for finite intersections (respectively finite sums) of subspaces.
Proposition 2.7. Let V be a vector space over F and S Ă V nonempty. The set
řn (
xSy “ i“1 αi vi : αi P F and vi P S, for all i “ 1, n, n P N is a vector subspace
over F of V .
The above vector space is called the vector space generated by S, or the linear hull
of the set S and is often denoted by spanpSq. It is the smallest subspace of V
which contains S, in the sense that for every U subspace of V with S Ă U it
follows that xSy Ă U.
Now we specialize the notion of sum of subspaces, to direct sum of subspaces.
Let V be a vector space over F and U be a subspace. On V one can define the
following binary relation RU : let u, v P V , u RU v iff u ´ v P U.
It can easily be verified that the relation RU is an equivalence relation, that is
RU rvs “ tu P V : v RU uu “ v ` U.
The quotient set (or factor set) V {RU is denoted by V {U and consists of the set of
all equivalence classes, that is
V {U “ tRU rvs : v P V u.
Proof. Indeed, let us define the sum of two equivalence class RU rvs and RU rws by
RU rvs ` RU rvs “ RU rv ` ws
Basis. Dimension. 34
The vector space from the previous theorem is called the factor vector space, or
the quotient vector space.
Example 2.12. Check whether the vectors p0, 1, 2q , p1, 2, 0q , p2, 0, 1q are linearly
independent in R3 .
holds.
Checking the above implication actually amounts (after computing the right hand
side) to investigating whether the linear system
$
’
’ α2 ` 2α2 “ 0
’
&
α1 ` 2α2 “ 0
’
’
’
% 2α
1 ` α2 “ 0
has only the trivial solution pα1 , α2 , α3 q “ p0, 0, 0q or not. But we can easily
compute the rank of the matrix, which is 3 due to
ˇ ˇ
ˇ ˇ
ˇ 0 1 2 ˇ
ˇ ˇ
ˇ ˇ
ˇ 1 2 0 ˇ “ ´9 ‰ 0,
ˇ ˇ
ˇ ˇ
ˇ 2 0 1 ˇ
Basis. Dimension. 36
to see that, indeed, the system has only the trivial solution, and hence the three
vectors are linearly independent.
We have the following theorem.
We will not prove this general theorem here, instead we will restrict to finite
dimensional vector spaces.
Theorem 2.14. Let V ‰ t0u be e finitely generated vector space over F. From
every finite system of generators one can extract a basis.
Proof. Let S “ tv1 , . . . , vr u be a finite generators system. It is clear that there are
nonzero vectors in S (otherwise V “ t0u). Let 0 ‰ v1 P S. The set tv1 u is linearly
independent (because αv1 “ 0 ñ α “ 0 from v1 ‰ 0). That means that S contains
linearly independent subsets. Now P pSq is finite (S being finite), and in a finite
number of steps we can extract a maximal linearly independent system, let say
B “ tv1 , . . . , vn u, 1 ď n ď r in the following way:
v2 P Szxv1 y,
v3 P Szxtv1 , v2 uy
..
.
Theorem 2.16. Let V be a finitely generated vector space over F. Every linearly
independent system of vectors L can be completed to a basis of V .
Proof. Let B “ te1 , . . . .en u be a basis of V , and let B1 te11 , . . . , e1m u a system of
vectors with m ą n. We show that B1 can not be a basis for V .
ř
Because B is a basis the vectors e1i can be uniquely written as e1i “ nj“1 aij ej ,
ř
1 ď i ď m. If B1 is linearly independent, then it follows that m 1
i“1 λi ei “ 0 implies
ř
λi “ 0, i “ 1, m, or, in other words, the system m i“1 aij λi “ 0, j “ 1, n has only
Definition 2.18. Let V ‰ t0u be a F vector space finitely generated. The number
of elements in a basis of V is called the dimension of V (it does not depend on the
Basis. Dimension. 38
choice of the basis, and it is denoted by dim F V ). The vector space V is said to be
of finite dimension. For V “ t0u , dim F V “ 0.
Remark 2.19. According to the proof of Theorem 2.17, if dim F V “ n then any
set of m ą n vectors is linear dependent.
Remark 2.21. The dimension of a finite dimensional vector space is equal to any
of the following:
Basis. Dimension. 39
In example 2.5 we have shown that S is a subspace of R3 . One can see that, from a
geometric point of view, S is a plane passing through the origin, so dim S “ 2.
This follows also from rewriting S as follows
(
S “ px, y, zq P R4 |x ` y ` z “ 0
“ tpx, y, ´x ´ yq |x, y P Ru
The vectors p1, 0, ´1q and p0, 1, ´1q are linearly independent so they form a basis
of S.
After each step B is still linearly independent. After n steps at most, the span of
B includes all the w’s. Thus B also spans V , and being linearly independent, it
follows that it is a basis.
V “ U ` W, and U X W “ t0u
v “ a1 u1 ` ¨ ¨ ¨ ` am um ` b1 w1 ` ¨ ¨ ¨ ` bn wm ,
v “ a1 u1 ` ¨ ¨ ¨ ` am um “ b1 w1 ` ¨ ¨ ¨ ` bn wm ,
Basis. Dimension. 41
so
a1 u1 ` ¨ ¨ ¨ ` am um ´ b1 w1 ´ ¨ ¨ ¨ ´ bn wm “ 0.
The next theorem relates the dimension of the sum and the intersection of two
subspaces with the dimension of the given subspaces:
dim pU ` W q “ m ` i ` j
“ pm ` iq ` pm ` jq ´ m
a1 u1 ` ¨ ¨ ¨ ` am um ` b1 v1 ` ¨ ¨ ¨ ` bi vi ` c1 w1 ` ¨ ¨ ¨ ` cj wj “ 0 .
We have
c1 w1 ` ¨ ¨ ¨ ` cj wj “ ´a1 u1 ´ ¨ ¨ ¨ ´ am um ´ b1 v1 ´ ¨ ¨ ¨ ´ bi vi
Basis. Dimension. 42
c1 w1 ` ¨ ¨ ¨ ` cj wj “ ´pd1 u1 ` ¨ ¨ ¨ ` dm um q .
a1 u1 ` ¨ ¨ ¨ ` am um ` b1 v1 ` ¨ ¨ ¨ ` bi vi “ 0 ,
so a’s and b’s are zero because the vectors tu1 , . . . , um , v1 . . . , vi u form a basis in U.
So all the a’s, b’s and c’s are zero, that means that
tu1 , . . . , um , v1 , . . . , vi , w1 , . . . , wj u are linearly independent, and because that
generates U ` W , they form a basis of U ` W .
The previous theorem shows that the dimension fits well with the direct sum of
spaces. That is, if U X W “ t0u, the sum is the direct sum and we have
This is true for the direct sum of any finite number of spaces as it is shown in the
next theorem:
and
dim V “ dim U1 ` ¨ ¨ ¨ ` dim Un .
Basis. Dimension. 43
Then
V “ U1 ‘ ¨ ¨ ¨ ‘ Un .
Proof. One can choose a basis for each Ui . By putting all these bases in one list,
we obtain a list of vectors which spans V (by the first property in the theorem),
and it is also a basis, because by the second property, the number of vectors in this
list is dim V .
Suppose that we have ui P Ui , i “ 1, n, such that
0 “ u1 ` ¨ ¨ ¨ ` un .
Every ui is represented as the sum of the vectors of basis of Ui , and because all
these bases form a basis of V , it follows that we have a linear combination of the
vectors of a base of V which is zero. So all the scalars are zero, that is all ui are
zero, so the sum is direct.
We end the section with two important observations. Let V be a vector space over
F (not necessary finite dimensional). Consider a basis B “ pei qiPI of V .
We have the first representation theorem:
Theorem 2.27. Let V be a vector space over F (not necessary finite dimensional).
Let us consider a basis B “ pei qiPI . For every v P V, v ‰ 0 there exist a unique
subset B1 Ď B, B1 “ tei1 , . . . , eik u and the nonzero scalars ai1 , . . . , aik P F˚ , such
that
ÿ
k
v“ aij eij “ ai1 ei1 ` ¨ ¨ ¨ ` aik eik .
j“1
ejs P tek1 , . . . , em u, for all s “ 1, n. Hence, m “ n and one may assume that
ÿ
n ÿ
n
v“ αji eji “ αki eki , αji ‰ 0, i “ 1, n, αki ‰ 0, i “ 1, n.
i“1 i“1
řn řn
Using the relation i“1 αji eji ´ i“1 αki eki “ 0 again we obtain that
αji “ αki , i P t1, . . . , nu, contradiction.
Example 2.28. Show that B “ tp1, 1q , p1, ´1qu.is a basis of R2 , and find the
representation with respect to B of the vector v “ p3, ´1q.
Our aim is to find the representation of v “ p3, ´1q with respect to B, that is, to
find two scalars x, y P R such that
Expressing the above equality component wise gives a system with two unknowns,
x and y $
& x ` y “ 3
% x ´ y “ ´1.
The scalars pa1 , . . . , an q are called the coordinates of the vector v in the basis B. It
1
is obvious that if we have another basis B , the coordinates of the same vector in
the new basis change. How we can measure this change? Let us start with a
situation that is a bit more general.
Theorem 2.29. Let V be a finite dimensional vector space over F with a basis
1 1
B “ te1 , . . . , en u. Consider the vectors S “ te1 , . . . , em u Ď V :
1
e1 “ a11 e1 ` ¨ ¨ ¨ ` a1n en
...
1
em “ am1 e1 ` ¨ ¨ ¨ ` amn en
Denote by A “ paij qi“1,m the matrix formed by the coefficients in the above
j“1,n
equations. The dimension of the subspace xSy is eqaul to the rank of the matrix A,
i.e. dimxSy “ rankA.
1 1
Consider now the case of m “ n in the above discussion. The set S “ te1 , . . . , en u
is a basis iff rankA “ n We have now
1
e1 “ a11 e1 ` ¨ ¨ ¨ ` a1n en
1
e2 “ a21 e1 ` ¨ ¨ ¨ ` a2n en
...
1
en “ an1 e1 ` ¨ ¨ ¨ ` ann en ,
1
representing the relations that change from the basis B to the new basis B “ S.
The matrix AJ is denoted by
¨ ˛
a11 a21 . . . an1
˚ ‹
˚ ‹
1 ˚ a12 a22 . . . an2 ‹
P pe,e q “˚
˚
‹
‹
˚ ... ... ... ... ‹
˝ ‚
a1n a2n . . . ann
.
The columns of this matrix are given by the coordinates of the vectors
1
of the new basis e with respect to the old basis e!
Remarks
1 1
• Consider the change of the basis from B to B with the matrix P pe,e q and
1 2 1 2
the change of the basis from B to B with the matrix P pe ,e q . We can think
Local computations 47
at the ”composition” of these two changes, i.e. the change of the basis from
2 2
B to B with the matrix P pe,e q . It is easy to see that one has
1 1 2 2
P pe,e q P pe ,e q “ P pe,e q
2
• If in the above discussion we consider B “ B one has
1 1
P pe,e q P pe ,eq “ In ,
that is
1 1
pP pe ,eq q´1 “ P pe,e q
1 1
v “ a1 e1 ` ¨ ¨ ¨ ` an en “ b1 e1 ` ¨ ¨ ¨ ` bn en ,
where pa1 , . . . an q and pb1 , . . . bn q are the coordinates of the same vector in the two
basis. We can write
¨ ˛ ¨ ˛
1
e e
˚ 1 ‹ ˚ 1 ‹
´ ˚ ‹
¯ ˚ e2 ‹ ´ ˚
¯ ˚ e1 ‹
‹
an ¨ ˚ ‹ “ b b ... b ¨˚ ‹.
2
pvq “ a1 a2 . . . ˚ ‹ 1 2 n ˚ ‹
˚ ... ‹ ˚ ... ‹
˝ ‚ ˝ ‚
1
en en
Local computations 48
Denote ¨ ˛
a
˚ 1 ‹
˚ ‹
˚ a2 ‹
pvqe “ ˚
˚
‹
‹
˚ ... ‹
˝ ‚
an
and ¨ ˛
b
˚ 1 ‹
˚ ‹
˚ b2 ‹
pvqe1 “ ˚
˚
‹.
‹
˚ ... ‹
˝ ‚
bn
the matrices of the coordinates of v in the two basis.
Denote further the basis columns
¨ ˛
e
˚ 1 ‹
˚ ‹
˚ e2 ‹
peq1n “˚
˚
‹
‹
˚ ... ‹
˝ ‚
en
or
1 1
pvqe1 “ pP pe,e q q´1 pvqe “ P pe ,eq pvqe .
Hence,
1
pvqe “ pP pe,e q qpvqe1 .
2.5 Problems
Problem 2.5.1. Show that for spanpv1 , . . . , vn q “ V one has
spanpv1 ´ v2 , v2 ´ v3 , . . . , vn´1 ´ vn , vn q “ V .
Problem 2.5.2. Find a basis for the subspace generated by the given vectors in
M3 pRq. ¨ ˛ ¨ ˛ ¨ ˛
1 2 0 ´1 2
3 0 1 2
˚ ‹ ˚ ‹ ˚ ‹
˚ ‹ ˚ ‹ ˚ ‹
˚2 4 1 ‹ ˚2 1 ´1‹ ,
, ˚´2 2 ´1‹ .
˝ ‚ ˝ ‚ ˝ ‚
3 1 ´1 0 1 1 ´1 2 1
Problem 2.5.3. Let V be a finite dimensional vector space dim V “ n. Show that
there exist one dimensional subspaces U1 , . . . , Un , such that
V “ U1 ‘ ¨ ¨ ¨ ‘ Un .
R2 “ U ‘ V “ V ‘ W “ W ‘ U.
Problem 2.5.7. Let U and W be subspaces of a vector space V and suppose that
each vector v P V has a unique expression of the form v “ u ` w where u belongs
to U and w to W. Prove that
V “ U ‘ W.
Problem 2.5.8. In Cra, bs find the dimension of the subspaces generated by the
following sets of vectors:
Problem 2.5.9. Find the dimension and a basis in the intersection and sum of
the following subspaces:
Problem 2.5.10. Let U, V, W be subspaces of some vector space and suppose that
U Ď W. Prove that
pU ` V q X W “ U ` pV X W q.
Problem 2.5.12. Let V, W be two vector spaces over the same field F. Find the
dimension and a basis of V ˆ W.
Problems 51
Problem 2.5.15. Let Mn pRq be the set of the real square matrices of order n,
and An , respectively Sn the set of symmetric, respectively skew-symmetric
matrices of order n. Show that Mn pRq “ An ‘ Sn .
Problem 2.5.16. Let us denote by Rn rXs the set of all polynomials having degree
at most n with real coefficients. Obviously Rn rXs is a subspace of RrXs with the
induced operations. Find the dimension of the quotient space Rn rXs{U where U is
the subspace of all real constant polynomials.
Problem 2.5.19. Find bases for the sum and intersection of the subspaces U and
W of R4 rXs generated by the respective sets of polynomials
t1 ` 2x ` x3 , 1 ´ x ´ x2 u and tx ` x2 ´ 3x3 , 2 ` 2x ´ 2x3 u.
3
Linear maps between vector spaces
Up to now we met with vector spaces. It is natural to ask about maps between
them, which are compatible with the linear structure of a vector space. These are
called linear maps, special maps which also transport the linear structure. They
are also called morphisms of vector spaces or linear transformations.
Definition 3.1. Let V and W be two vector spaces over the same field F. A linear
map from V to W is a map f : V Ñ W which has the property that
f pαv ` βuq “ αf pvq ` βf puq for all v, u P V and α, β P F.
We shall define now two important notions related to a linear map, the kernel and
the image.
Consider the sets:
53
54
Definition 3.2. The sets ker f and f pV q are called the kernel (or the null space),
respectively the image of f .
Proposition 3.3. The kernel and the image of a linear map f : V Ñ W are
subspaces of V and W respectively.
By definition
(
ker T “ px, yq P R2 |T px, yq “ p0, 0q
(
“ px, yq P R2 | px ` y, x ` yq “ p0, 0q
(
“ px, yq P R2 |x ` y “ 0 .
For the finite dimensional case the dimension of ker and im of a linear map
between vector spaces are related by the following:
ÿ
n
αk f pek q “ 0W ,
k“m`1
and by the linearity of f ,
ÿ
n
fp αk ek q “ 0W .
k“m`1
Hence
ÿ
n
1
v “ αk ek P ker f
k“m`1
and v 1 can be written in terms of e1 , . . . , em . This is only compatible with the fact
that e1 , . . . , en form a basis of V if αm`1 “ ¨ ¨ ¨ “ αn “ 0, which implies the linear
independence of the vectors f pem`1 q, . . . , f pen q.
The fact that f pV q “ W implies that ker f “ t0V u follows by reversing the
arguments.
p : V Ñ V1 , ppvq “ v1 , @v P V
The next proposition shows that the kernel and the image of a linear map
characterize the injectivity and surjectivity properties of the map.
2. f is onto (surjective) ðñ f pV q “ W .
Next we shall study how special maps act on special systems of vectors.
řn
αi P F, i “ 1, n such that i“1 αi vi “ v. It follows that
ÿ
n ÿ
n
w “ f pvq “ f p αi vi q “ αi f pvi q.
i“1 i“1
3. Because f is bijective and S is a basis for V , it follows that both 1. and 2. hold,
that is f pSq is a basis for W .
Proposition 3.14. With the operations defined above LpV, W q becomes a vector
space over F.
EndpLq “ tf : V Ñ V | f linear u.
“ αhpv1 q ` βhpv2 q.
The last statement follows from the fact that h is a linear bijection.
It can be shown that the composition is distributive with respect to the sum of
linear maps, so EndpV q becomes an unitary ring.
It can easily be realized that:
The group of automorphisms of a linear space is called the general linear group
and is denoted by GLpV q.
p2 “ p,
1. ker p ‘ ppV q “ V
v “ v1 ` v2 , @ v P V,
1
ps : V Ñ V, ps pvq “ pv ` spvqq
2
1
qs : V Ñ V, qs pvq “ pv ´ spvqq
2
Example 3.21. Let V be a vector space and f : V Ñ V a linear map such that
ker f “ imf . Determine the set imf 2 , where f 2 denotes the composition of f with
itself, f 2 “ f ˝ f .
imf 2 “ imf ˝ f
“ f ˝ f pV q
“ f pf pV qq .
But, f pV q “ imf “ ker f is the set of all vectors which are mapped by f to zero, so
imf 2 “ f pker f q
“ 0.
Local form of a linear map 63
Now we want to see how the image of a vector by a linear map can we expressed.
ř
Let v P V, v “ m J
i“1 vi ei , or in the matrix notation pvqe peq1m , where, as usual
¨ ˛
v1
˚ ‹
˚ ‹
˚ v2 ‹
pvqe “ ˚
˚ ..
‹
‹
˚ . ‹
˝ ‚
vn
and ¨ ˛
e1
˚ ‹
˚ ‹
˚ e2 ‹
peq1m “˚
˚ ..
‹.
‹
˚ . ‹
˝ ‚
em
řn
Now denote T pvq “ w “ j“1 wj ej P W , we have
T pvq “ pwqJ
f pf q1n .
řm
T being linear, we have T pvq “ i“1 vi T pei q, or, again in matrix notation:
T pvq “ pvqJ
e pT peqq1m .
pf,eq
From the definition of MT it follows that
pf,eq J
pT peqq1m “ pMT q pf q1n .
So finally we have
pf,eq J
pwqJ J
f pf q1n “ pvqe pMT q pf q1n .
¨ ˛
´3 0 2
˚ ‹
3 3 ˚ ‹
Example 3.23. Let T : R Ñ R , T “ ˚ 1 1 0 ‹. Find a basis in ker T and
˝ ‚
´2 1 2
3
find the dimension of T pR q .
the matrix of the system being exactly T . To solve this system we need to
compute the rank of the matrix T . We get that
ˇ ˇ
ˇ ˇ
ˇ ´3 0 2 ˇ
ˇ ˇ
ˇ ˇ
ˇ 1 1 0 ˇ“0
ˇ ˇ
ˇ ˇ
ˇ ´2 1 2 ˇ
3
x “ α, y “ ´x, z “ x.
2
` ˘
and hence, a basis in ker T consists only of 1, ´1, 32 , dim ker T “ 1.
Based on the dimension formula
` ˘
dim ker T ` dim T R3 “ dim R3
MS˝T “ Ms MT .
Proof. Indeed, one can easily see that for v P V we have pT pvqq “ MT pvq where
pT pvqq, respectively pvq stand for the coordinate of T pvq, respectively v in the
appropriate bases. Similarly, for w P W one has pSpwqq “ MS pwq.
Hence, pS ˝ T pvqq “ pSpT pvqqq “ MS pT pvqq “ MS MT pvq, or, equivalently
MS˝T “ MS MT .
pf 1 ,e1 q 1 pf,eq 1
Theorem 3.25. In the above conditions MT “ P pf ,f q MT P pe,e q .
Problems 67
pf 1 ,e1 q pf 1 ,e1 q 1
pwqf 1 “ MT pvqe1 “ MT P pe ,eq pvqe .
1 1 1 pf,eq
pwqf 1 “ P pf ,f q pwqf “ P pf ,f q pT pvqqf “ P pf ,f q MT pvqe .
1 1
Taking into account that pP pe ,eq q´1 “ P pe,e q we get
3.3 Problems
Problem 3.3.1. Consider the following mappings T : R3 Ñ R3 . Study which one
of them is a linear mapping.
b) T px1 , x2 , x3 q “ px3 , x1 , x2 q.
c) T px1 , x2 , x3 q “ px1 ´ 1, x2 , x3 q.
d) T px1 , x2 , x3 q “ px1 ` x2 , x2 ´ x3 , x1 ` x2 ` x3 q.
respectively ¨ ˛
´1 4 2
˚ ‹
˚ ‹
MT2 “ ˚ 0 4 1‹
˝ ‚
0 0 5
in the canonical basis of R3 .
Problem 3.3.4. Let V be a complex vector space and let T P EndpV q. Show that
there exists a basis in V such that the matrix of T relative to this basis is upper
triangular.
Problems 69
Problem 3.3.9. Find all the linear mappings T : R Ñ R with the property
im T “ ker T .
Find all n P N such that there exists a linear mapping T : Rn Ñ Rn with the
property im T “ ker T .
V { ker T » im T.
pU ` W q{W » U{pU X W q.
a) te1 , e3 , e2 , e4 u.
b) te1 , e1 ` e2 , e1 ` e2 ` e3 , e1 ` e2 ` e3 ` e4 u.
c) te4 ´ e1 , e3 ` e4 , e2 ´ e4 , e4 u.
à
m
V “ Ui ,
i“1
72
Invariant subspaces. Proper vectors and values 73
However we have a problem: if we want to apply tools which are commonly used in
the theory of linear maps (such as taking powers for example) the problem is that
generally T may not map Uj into itself, in other words T |Uj may not be an
operator on Uj . For this reason it is natural to consider only that kind of
decomposition for which T maps every Uj into itself.
Of course that another natural question arises when dealing with invariant
subspaces. How does an operator behave on an invariant subspace of dimension
one? Every one dimensional subspace is of the form U “ tλu|λ P Fu. If U is
invariant by T it follows that T puq should be in U, and hence there should exist a
scalar λ P F such that T puq “ λu. Conversely if a nonzero vector u exists in V
such that T puq “ λu, for some λ P F, then the subspace U spanned by u is
invariant under T and for every vector v in U one has T pvq “ λv. It seems
reasonable to give the following definition:
Definition 4.2. Let T P EndpV q be an operator on a vector space over the field F.
A scalar λ P F is called eigenvalue (or proper value) for T if there exists a nonzero
vector v P V such that T pvq “ λv. A corresponding vector satisfying the above
equality is called eigenvector (or proper vector) associated to the eigenvalue λ.
detpMT ´ λIn q “ 0.
1
detpP ´1 pMT ´ λIqP q “ detpMT ´ λIq detpP q “ detpMT ´ λIq,
detpP q
which proves our claim.
vk P spantv1 , . . . , vk´1 u.
vk “ a1 v1 ` . . . ak´1 vk´1 .
Invariant subspaces. Proper vectors and values 75
It follows that
The linear maps which have exactly n “ dim V linearly independent eigenvectors
have very nice and simple properties. This is the happiest case we can meet with
in the class of linear maps.
Recall that matrices A and B are similar if there is an invertible matrix P such
that B “ P AP ´1. Hence, a matrix A is diagonalizable if it is similar to a diagonal
matrix D.
The minimal polynomial of an operator 76
T m T n “ T m`n , for m, n P Z.
ppzq “ a0 ` a1 z ` . . . am z m , z P F
ppT q “ a0 I ` a1 T ` . . . am T m .
This is a new use of the same symbol p, because we are applying it to operators
not only to elements in F. If we fix the operator T we obtain a function defined on
FrXs with values in EndpV q, given by p Ñ ppT q which is linear. For p, q P FrXs we
define the operator pq given by ppqqpT q “ ppT qqpT q.
Now we begin the study of the existence of eigenvalues and of their properties.
Theorem 4.6. Every operator over a finite dimensional, nonzero, complex vector
space has an eigenvalue.
The minimal polynomial of an operator 77
This set is a linearly dependent system of vectors (they are n ` 1) vectors and
dim V “ n. Then there exist complex numbers, a0 , . . . an , not all 0, such that
0 “ a0 v ` a1 T pvq ` ¨ ¨ ¨ ` an T n pvq .
Let m be the largest index such that am ‰ 0. Then we have the decomposition
a0 ` a1 z ` ¨ ¨ ¨ ` am z m “ a0 pz ´ λ1 q . . . pz ´ λm q .
It follows that
0 “ a0 v ` a1 T pvq ` . . . an T n pvq
“ pa0 I ` a1 T ` . . . an T n qpvq
“ a0 pT ´ λ1 Iq . . . pT ´ λm Iqpvq .
which means that T ´ λj I is not injective for at least one j, or equivalently T has
an eigenvalue.
Remark 4.7. The analogous statement is not true for real vector spaces. But on
real vector spaces there are always invariant subspaces of dimension 1 or 2.
Theorem 4.9. Every operator on an odd dimensional real vector space has an
eigenvalue.
Proof. Let T P EndpV q and n “ dim V odd. The eigenvalues of T are the roots of
the characteristic polynomial that is detpMT ´ λIn q. This polynomial is a
polynomial of degree n in λ, hence, since n is odd, the equation detpMT ´ λIn q “ 0
has at least one real solution.
A central goal of linear algebra is to show that a given operator T P EndpV q has a
reasonably simple matrix in a given basis. It is natural to think that reasonably
simple means that the matrix has as many 01 s as possible.
Recall that for a basis tek , k “ 1, nu,
ÿ
n
T pek q “ aik ei ,
i“1
Proof. 1ô2 obviously follows from a moment’s tought and the definition. Again
3ñ2. It remains only to prove that 2ñ3.
So, suppose 2 holds. Fix k P t1, . . . , nu. From 2 we have
T pvq P spante1 , . . . , ek u,
consequently 3. holds.
Theorem 4.11. Suppose that V is a complex vector space and T P EndpV q. Then
there exists a basis of V such that T is an upper-triangular matrix with respect to
this basis.
One of the good points of this theorem is that, if we have this kind of basis, we can
decide if the operator is invertible by analysing the matrix of the operator.
Theorem 4.12. Suppose T P EndpV q has an upper triangular matrix with respect
to some basis of V . Then T is invertible if and only if all the entries on the
diagonal are non zero.
Proof. Let te1 , . . . , en u be a basis of V with respect to which T has the matrix
¨ ˛
λ1 . . . ˚
˚ ‹
˚ ‹
˚ 0 λ2 ... ‹
MT “ ˚
˚
‹
‹
˚ 0 0 ... ‹
˝ ‚
0 0 0 λn
The minimal polynomial of an operator 81
We will prove that T is not invertible iff one of the λk ’s equals zero. If λ1 “ 0, then
T pv1 q “ 0, so T is not invertible, as desired.
Suppose λk “ 0, 1 ă k ď n. The operator T maps the vectors e1 , . . . , ek´1 in
spante1 , . . . , ek´1 u and, because λk “ 0, T pek q P te1 , . . . , ek´1 u. So, the vectors
T pe1 q, . . . , T pek q are linearly dependent (they are k vectors in a k ´ 1 dimensional
vector space, spante1 , . . . , ek´1 u. Consequently T is not injective, and not
invertible.
Suppose that T is not invertible. Then ker T ‰ t0u, so v P V, v ‰ 0 exists such
that T pvq “ 0. Let
v “ a1 e1 ` ¨ ¨ ¨ ` an en
v “ a1 e1 ` ¨ ¨ ¨ ` ak ek ,
and
0 “ T pvq,
0 “ T pa1 e1 ` ¨ ¨ ¨ ` ak ek q,
0 “ pa1 T pe1 q ` ¨ ¨ ¨ ` ak´1 T pek´1 qq ` ak T pek q.
The term pa1 T pe1 q ` ¨ ¨ ¨ ` ak´1 T pek´1 qq is in spante1 , . . . , ek´1 u, because of the
form of MT . Finally T pek q P spante1 . . . , ek´1 u. Thus when T pek q is written as a
linear combination of the basis te1 , . . . , en u, the coefficient of ek will be zero. In
other words, λk “ 0.
Theorem 4.13. Suppose that T P EndpV q has an upper triangular matrix with
respect to some basis of V . Then the eigenvalues of T are exactly of the entries on
the diagonal of the upper triangular matrix.
Proof. Suppose that we have a basis te1 , . . . , en u such that the matrix of T is
upper triangular in this basis. Let λ P F, and consider the operator T ´ λI. It has
Diagonal matrices 82
the same matrix, except that on the diagonal the entries are λi ´ λ if those in the
matrix of T are λj . It follows that T ´ λI is not invertible iff λ is equal with some
λj . So λ is proper value as desired.
The next proposition imposes several conditions on an operator that are equivalent
to having a diagonal matrix.
Suppose 2 holds, then V has a basis consisting of eigenvectors. Then every vector
in V is a linear combination of eigenvectors of T , that is
0 “ u1 ` ¨ ¨ ¨ ` un ,
Let te11 , . . . , e1i1 u, . . . , ten1 , . . . , enin u bases in kerpT ´ λ1 Iq, . . . , kerpT ´ λn Iq. Then
dim V “ i1 ` ¨ ¨ ¨ ` in , and te11 , . . . , e1i1 , . . . , en1 , . . . , enin u are linearly independent.
Hence V “ spante11 , . . . , e1i1 , . . . , en1 , . . . , enin u which shows that 2 holds.
¨ ˛
2 ´1 ´1
˚ ‹
˚ ‹
Example 4.16. Consider the matrix A “ ˚´1 2 ´1‹ . Show that A is
˝ ‚
´1 ´1 0
diagonalizable and find the diagonal matrix similar to A.
respectively
tpα, ´α, 0q : α P Ru.
¨ ˛
1 1 2
˚ ‹
˚ ‹
We have P ´1 “ 61 ˚2 2 ´2‹ .
˝ ‚
3 ´3 0
Hence, the diagonal matrix similar to A is
¨ ˛
´1 0 0
˚ ‹
˚ ‹
D “ P ´1AP “ ˚ 0 2 0‹ .
˝ ‚
0 0 3
Obviously one may directly compute D, by knowing, that D is the diagonal matrix
having the eigenvalues of A on its main diagonal.
3. T ´1 pvq “ u such that T puq “ v. But v “ λ´1 T pvq “ T pλ´1 vq, hence
T puq “ T pλ´1 vq. Since T is injective we have u “ λ´1 v, or equivalently
T ´1 pvq “ λ´1 v.
or, equivalently
` ˘
T 2 ` T ` I v “ 0.
Now, we apply the linear map T ´ I (recall that the linear maps form a vector
space, so the sum or difference of two linear maps is still linear) to the above
relation to get
` ˘
pT ´ Iq T 2 ` T ` I v “ 0.
as desired.
Let V be a vector space of finite dimension n over a field F. Let T : V Ñ V and let
λ0 be an eigenvalue of T . Consider the matrix form of the endomorphism in a
given basis, T pvq “ MT v. The eigenvalues are the roots of the characteristic
polynomial detpMt ´ λIn q “ 0. It can be proved that this polynomial does not
depend on the basis and of the matrix MT . So, it will be called the characteristic
polynomial of the endomorphism T , and it will be denoted by P pλq, and of course
deg P “ n. Sometimes it is called the characteristic polynomial of the matrix, but
we understand that is the matrix associated to an operator.
Denote by mpλ0 q the multiplicity of λ0 as a root of this polynomial. Associated to
the proper value λ0 we consider the proper subspace corresponding to λ0 :
Consider a basis of V and let MT be the matrix of T with respect to this basis. We
havev that:
1
detpP ´1 pMT ´ λIqP q “ detpMT ´ λIq detpP q “ detpMT ´ λIq.
detpP q
But since the first few columns of P ´1 MT P are diagonal with λ0 on the diagonal
we have that the characteristic polynomial of P ´1MT P has a factor of at least
pλ0 ´ λqr , so the algebraic multiplicity of λ0 is at least r.
The value dim Epλ0 q is called the geometric multiplicity of the eigenvalue λ0 .
Let T P EndpV q, and suppose that the roots of the characteristic polynomial are in
F. Let λ be a root of the characteristic polynomial, i.e. an eigenvalue of T .
Consider m the algebraic multiplicity of λ and q “ dim Epλq, the geometric
multiplicity of λ.
It is possible to find q eigenvectors and m ´ q principal vectors (also called
generalized eigenvectors), all of them linearly independent, and an eigenvector v
and the corresponding principal vectors u1 , . . . , ur satisfy
matrix of T relative to a Jordan basis is called a Jordan matrix, and it has the form
¨ ˛
J
˚ 1 ‹
˚ ‹
˚ J2 ‹
˚ ‹
˚ .. ‹
˚ . ‹
˝ ‚
Jp
The J’s are matrices, called Jordan cells. Each cell represents the contribution of
an eigenvector v, and the corresponding principal vectors, u1 , . . . ur , and it has the
form
¨ ˛
λ 1
˚ ‹
˚ ‹
˚ λ 1 ‹
˚ ‹
˚ ‹
˚ λ 1 ‹ P Mr`1 pFq
˚ ‹
˚ .. ‹
˚ . 1 ‹
˝ ‚
λ
It is easy to see that the Jordan matrix is a diagonal matrix iff there are no
principal vectors iff mpλq “ dim Epλq for each eigenvalue λ.
Let MT be the matrix of T with respect to a given basis B, and J be the Jordan
matrix with respect to a Jordan basis B 1 . Late P be the transition matrix from B
to B 1 , hence it have columns consisting of either eigenvectors or generalized
eigenvectors. Then J “ P ´1MT P , hence MT “ P JP ´1 .
The solutions of this system lie in the set tpα, 2α ` 1, βq : α, β P Ru, hence a
generalized eigenvector, is u1 “ p1, 3, 0q.
Note that v1 , u1 , v¨
2 are linear
˛ independent, ¨
hence we take
˛ the transition matrix
1 1 0 3 ´1 0
˚ ‹ ˚ ‹
˚ ‹ ˚ ‹
P “ rv1 |u1 |v2 s “ ˚2 3 0‹ . Then P ´1 “ ˚´2 1 0‹ , hence
˝ ‚ ˝ ‚
1 0 1 ´3 1 1
¨ ˛
2 1 0
˚ ‹
˚ ‹
J “ P ´1 AP “ ˚0 2 0‹ .
˝ ‚
0 0 2
Ep2q “ kerpA ` 2Iq “ tp5α, 3α, ´7αq : α P Ru. Hence the dimension of Ep2q is 1,
consequently the eigenvalue λ “ 2 has geometric multiplicity 1. Therefore we can
take the linear independent eigenvector v “ p5, 3, ´7q. Note that we need two
generalized eigenvectors, which can be obtained as a solution of the system
pA ` 2Iqu1 “ v,
respectively
pA ` 2Iqu2 “ u1 .
The solutions of the first system lie in the set tp´ 1`5α
7
, ´ 2`3α
7
, αq : α P Ru, hence a
generalized eigenvector, for α “ 4 is u1 “ p´3, ´2, 4q.
The solutions of the system pA ` 2Iqu2 “ u1 with u1 “ p´3, ´2, 4q lie in the set
tp´ 3`5α
7
, 1´3α
7
, αq : α P Ru, hence a generalized eigenvector, for α “ 5 is
u1 “ p´4, ´2, 5q. Note that v, u1 , u2¨are linear independent,
˛ hence we take the
5 ´3 ´4
˚ ‹
˚ ‹
transition matrix P “ rv1 |u1 |u2 s “ ˚ 3 ´2 ´2‹ . Then
˝ ‚
´7 4 5
¨ ˛
´2 ´1 ´2
˚ ‹
˚ ‹
P ´1 “ ˚´1 ´3 ´2‹ , hence
˝ ‚
´2 1 ´1
¨ ˛
´2 1 0
˚ ‹
´1 ˚ ‹
J “ P AP “ ˚ 0 ´2 1 ‹ .
˝ ‚
0 0 ´2
4.5 Problems
Problem 4.5.1. Find the eigenvalues and eigenvectors of the operator
f 1 pxq
T : C 8 p1, bq Ñ C 8 p1, bq, T pf qpxq “ xex2
.
Problems 92
¨ ˛
1 5
Problem 4.5.2. Find matrices which diagonalize the following: aq ˝ ‚.
3 3
¨ ˛
1 2 ´1
˚ ‹
˚ ‹
bq ˚1 0 1 ‹.
˝ ‚
4 ´4 5
Problem 4.5.3. Find the Jordan canonical form and the transition matrix for the
matrix ¨ ˛
2 1 ´1
˚ ‹
˚ ‹
˚3 ´2 3 ‹ .
˝ ‚
2 ´2 3
Problem 4.5.4. Prove that a square matrix and its transpose have the same
eigenvalues.
Problem 4.5.5. Find the Jordan canonical form and the transition matrix for the
matrix ¨ ˛
6 6 ´15
˚ ‹
˚ ‹
˚1 5 ´5 ‹ .
˝ ‚
1 2 ´2
Problem 4.5.7. Find the Jordan canonical form and the transition matrix for the
matrix ¨ ˛
4 1 1
˚ ‹
˚ ‹
˚´2 2 ´2‹ .
˝ ‚
1 1 4
Problems 93
Problem 4.5.9. Find the Jordan canonical form and the transition matrix for the
matrix ¨ ˛
7 ´12 6
˚ ‹
˚ ‹
˚10 ´19 10‹ .
˝ ‚
12 ´24 13
Problem 4.5.12. Find the Jordan canonical form and the transition matrix for
the matrix ¨ ˛
4 ´5 2
˚ ‹
˚ ‹
˚5 ´7 3‹ .
˝ ‚
6 ´9 4
Problem 4.5.14. Find the Jordan canonical form and the transition matrix for
the matrix ¨ ˛
1 1 0
˚ ‹
˚ ‹
˚´4 ´2 1 ‹ .
˝ ‚
4 1 ´2
Problems 94
Problem 4.5.16. Find the Jordan canonical form and the transition matrix for
the matrices ¨ ¨˛ ˛
1 ´3 3 4 6 ´15
˚ ‹ ˚ ‹
˚ ‹ ˚ ‹
˚´2 ´6 13‹ , ˚1 3 ´5 ‹ .
˝ ‚ ˝ ‚
´1 ´4 8 1 2 ´4
Problem 4.5.18. Find the Jordan canonical form and the transition matrix for
the matrix ¨ ˛
2 6 ´15
˚ ‹
˚ ‹
˚1 1 ´5 ‹ .
˝ ‚
1 2 ´6
5
Inner product spaces
95
Basic definitions and results 96
An inner product space is a pair pV, x¨, ¨yq, where V is vector space and x¨, ¨y is an
inner product on V .
xv, wy “ v1 w 1 ` ¨ ¨ ¨ ` vn w n .
This is the typical example of an inner product, called the Euclidean inner
product, and when Fn is referred to as an inner product space, one should assume
that the inner product is the Euclidean one, unless explicitly stated otherwise.
¨ ˛
a b
Example 5.2. Let A P M2 pRq, A “ ˝ ‚ be a positive definite matrix, that
b c
2
is a ą 0, detpAq ą ˛ for every u “ pu1 , u2q, v “ pv1 , v2 q P R we define
¨0. Then
u1
xu, vy “ pv1 v2 qA ˝ ‚.
u2
It can easily be verified that x¨, ¨y is an inner product on the real linear space R2 .
If A “ I2 we obtain the usual inner product xu, vy “ u1 v1 ` u2 v2 .
From the definition one can easily deduce the following properties of an inner
product:
xv, 0y “ x0, vy “ 0,
}¨}:V ÑR
A normed space is a pair pV, } ¨ }q, where V is a vector space and } ¨ } is a norm on
V.
Example 5.4. On the real linear space Rn one can define a norm in several ways.
Indeed, for any x “ px1 , x2 , . . . , xn q P Rn define its norm as
a
}x} “ x21 ` x22 ` ¨ ¨ ¨ ` x2n . One can easily verify that the axioms in the definition
of norm are satisfied. This norm is called the euclidian norm.
More generally, for any p P R, p ě 1 we can define
1
}x} “ p|x1 |p ` |x2 |p ` ¨ ¨ ¨ ` |xn |p q p , the so called p´norm on Rn .
Another way to define a norm on Rn is }x} “ maxt|x1 |, |x2 |, . . . , |xn |u. This is the
so called maximum norm.
In this course we are mainly interested in the inner product spaces. But we should
a
point out that an inner product on V defines a norm, by }v} “ xv, vy for v P V ,
and a norm on V defines a metric by dpv, wq “ }w ´ v}, for v, w P V .
On the other hand, from their generality point of view the metrics are the most
general ones (can be defined on any set), followed by norms (which assumes the
linearity of the space where is defined) and on the last position is the inner
product. It should be pointed that every inner product generates a norm, but not
every norm comes from an inner product, as is the case for the max norm defined
above.
For an inner product space pV, x¨, ¨yq the following identity is true:
C G
ÿ
m ÿ
n ÿ
m ÿ
n
αi vi , βj wj “ αi β j xvi , wj y.
i“1 j“1 i“1 j“1
In a real inner product space we can define the angle of two vectors as
{ xv, wy
pv, wq “ arccos
}v} ¨ }w}
Basic definitions and results 99
We have
{ π
vKw ô xv, wy “ 0 ô pv, wq “ .
2
Theorem 5.8. (Parallelogram law) Let V be an inner product space and
u, v P V . Then
Proof.
“ 2p}u}2 ` }v}2q.
Proof.
}u ` v}2 “ xu ` v, u ` vy
“ }u}2 ` }v}2 .
Now we are going to prove one of the most important inequalities in mathematics,
namely the Cauchy-Schwartz inequality. There are several methods of proof for
this, we will give one related to our aims.
Basic definitions and results 100
ˆ ˙
xu, vy xu, vy
u“ v` u´ v .
}v}2 }v}2
Theorem 5.10. Cauchy-Schwartz Inequality Let V be an inner product space
and u, v P V . Then
|xu, vy| ď }u} ¨ }v}.
The equality holds iff one of u, v is a scalar multiple of the other (u and v are
collinear).
Proof. Let u, v P V . If v “ 0 both sides of the inequality are 0 and the desired
´ ¯
xu,vy xu,vy
result holds. Suppose that v ‰ 0. Write u “ }v}2 v ` u ´ }v}2 v . Taking into
xu,vy xu,vy
account that the vectors }v}2
v and u ´ }v}2
v are orthogonal, by the Pythagorean
theorem we obtain
› › › ›2
› xu, vy ›2 › xu, vy ›
}u} 2
“ ›› v › ` ›u ´ v ›
}v}2 › › }v}2 ›
› ›2
|xu, vy|2 ›› xu, vy ››
“ ` ›u ´ v
}v}2 }v}2 ›
|xu, vy|2
ě ,
}v}2
One of the reason that one studies orthonormal families is that in such special
bases the computations are much more simple.
for all α1 , α2 , . . . , αm P F.
α1 e1 ` α2 e2 ` ¨ ¨ ¨ ` αm em “ 0.
such that
spanpv1 , v2 , . . . , vk q “ spanpe1 , e2 . . . , ek q
Consider
B F
vj ´ xvj , e1 ye1 ´ ¨ ¨ ¨ ´ xvj , ej´1yej´1
xej , ek y “ , ek
}vj ´ xvj , e1 ye1 ´ ¨ ¨ ¨ ´ xvj , ej´1yej´1 }
xvj , ek y ´ xvj , ek y
“
}vj ´ xvj , e1 ye1 ´ ¨ ¨ ¨ ´ xvj , ej´1 yej´1}
“ 0,
Both lists being linearly independent (the first one by hypothesis and the second
one by orthonormality), it follows that the generated subspaces above have the
same dimension j, so they are equal.
tv1 “ p0, 1, 1, 0q, v2 “ p0, 4, 0, 1q, v3 “ p1, ´1, 1, 0q, v4 “ p1, 3, 0, 1qu.
xv2 , u1y 4
u2 “ v2 ´ u1 “ p0, 4, 0, 1q ´ p0, 1, 1, 0q “ p0, 2, ´2, 1q.
xu1 , u1 y 2
ˆ ˙
xv3 , u1 y xv3 , u2 y 1 1 4
u3 “ v3 ´ u1 ´ u2 “ 1, ´ , , .
xu1, u1 y xu2, u2 y 9 9 9
ˆ ˙
xv4 , u1 y xv4 , u2y xv4 , u3y 1 1 1 2
u4 “ v4 ´ u1 ´ u2 ´ u3 “ , ,´ ,´ .
xu1, u1 y xu2 , u2 y xu3 , u3 y 11 22 22 11
It can easily be verified that the list tu1 , u2, u3 , u4 u is orthogonal. Take now
ui
wi “ }ui }
, i “ 1, 4. We obtain
ˆ ˙
1 1
w1 “ 0, ? , ? , 0 ,
2 2
ˆ ˙
2 2 1
w2 “ 0, , ´ , ,
3 3 3
ˆ ˙
3 1 1 4
w3 “ ? , ´ ? , ? , ? ,
11 3 11 3 11 3 11
ˆ? ? ? ? ˙
22 22 22 2 22
w4 “ , ,´ ,´ .
11 22 22 11
Obviously the list tw1 , w2 , w3 , w4 u is orthonormal.
Now we can state the main results in this section.
Orthogonal complement 105
The next proposition shows that any orthonormal list can be extended to an
orthonormal basis.
Corollary 5.21. Suppose that V is a complex vector space and T P EndpV q. Then
T has an upper triangular form with respect to some orthonormal basis of V .
vector in U i.e.:
U K “ tv P V |xv, uy “ 0, @u P Uu.
V “ U ‘ UK
V “ U ` UK
Denote the first vector by u and the second by w. Clearly u P U. For each
j P t1, 2, . . . , mu one has
“ 0
a) U1 “ pU1K qK .
Proof. a) We show first that U1 Ď pU1K qK . Let u1 P U1 . Then for all v P U1K one has
vKu1 . In other words xu1 , vy “ 0 for all v P U1K . Hence u1 P pU1K qK .
Assume now that pU1K qK Ę U1 . Hence, there exists u2 P pU1K qK zU1 . Since
V “ U1 ‘ U1K we obtain that there exists u1 P U1 such that u2 ´ u1 P U1K p˚q.
On the other hand, according to the first part of proof u1 P pU1K qK and pU1K qK is a
linear subspace, hence u2 ´ u1 P pU1K qK . Hence, for all v P U1K we have
pu2 ´ u1 qKv p˚˚q.
p˚q and p˚˚q implies that pu2 ´ u1 qKpu2 ´ u1 q that is xu2 ´ u1 , u2 ´ u1 y “ 0, which
leads to u1 “ u2 contradiction.
b) For v P pU1 ` U2 qK one has xv, u1 ` u2 y “ 0 for all u1 ` u2 P U1 ` U2 . By taking
u2 “ 0 we obtain that v P U1K and by taking u1 “ 0 we obtain that v P U2K . Hence
pU1 ` U2 qK Ď U1K X U2K .
Conversely, let v P U1K X U2K . Then xv, u1y “ 0 for all u1 P U1 and xv, u2y “ 0 for all
u2 P U2 . Hence xv, u1 ` u2y “ 0 for all u1 P U1 and u2 P U2 , that is v P pU1 ` U2 qK .
c) According to a) ppU1 X U2 qK qK “ U1 X U2 .
According to b) and a) pU1K ` U2K qK “ pU1K qK X pU2K qK “ U1 X U2 .
Hence, ppU1 X U2 qK qK “ pU1K ` U2K qK which leads to pU1 X U2 qK “ U1K ` U2K .
“ tpx1 , x2 , x3 , x1 ´ x2 ` x3 q |x1 , x2 , x3 P Ru
The three vectors p1, 0, 0, 1q , p0, 1, 0, ´1q , p0, 0, 1, 1q are linearly independent (the
rank of the matrix they form is 3), so they form a basis of U and dim U “ 3.
The dimension formula
dim U ` dim U K “ dim R4
• if v0 P VL then L “ VL .
• v0 P L because v0 “ v0 ` 0 P v0 ` VL .
• for v1 , v2 P L we have v1 ´ v2 P VL .
3. Two linear manifolds L1 and L2 are called parallel if VL1 Ă VL2 or VL2 Ă VL1 .
L “ tv “ v0 ` α1 v1 ` ¨ ¨ ¨ ` αk vk |αi P F, i “ 1, ku
$
’
’
’
’ x1 “ x01 ` α1 x11 ` ¨ ¨ ¨ ` αk x1k
’
&
..
’ .
’
’
’
’
%xn “ x0n ` α1 xn1 ` ¨ ¨ ¨ ` αk xnk
The rank of the matrix pxij qi“1,n is k because the vectors v1 , . . . , vk are linearly
j“1,k
independent.
It is worthwhile to mention that:
Theorem 5.28. Let us consider V an n-dimensional vector space over the field F.
Then any subspace of V is the kernel of a surjective linear map.
Remark 5.29. In fact the map constructed in the previous theorem is nothing
but the projection on U parallel to the space spante1 , . . . , ek u.
Theorem 5.30. Let V, U two linear spaces over the same field F. If T : V Ñ U is
a surjective linear map, then for every u0 P U, the set L “ tv P V |T pvq “ u0 u is a
linear manifold.
Proof. T being surjective, there exists v0 P V with T pv0 q “ u0 . We will show that
tv ´ v0 |v P Lu “ ker T .
Let v P L. We have T pv ´ v0 q “ T pvq ´ T pv0 q “ 0, so tv ´ v0 |v P Lu Ď ker T .
Let v1 P ker T , i.e. T pv1 q “ 0. Write v1 “ pv1 ` v0 q ´ v0 . T pv1 ` v0 q “ u0, so
pv1 ` v0 q P L. Hence, v1 P tv ´ v0 |v P Lu or, in other words ker T Ď tv ´ v0 |v P Lu.
Consequently L “ v0 ` ker T, which shows that L is a linear manifold.
Theorem 5.31. Let V a linear space of dimension n. Then, for every linear
manifold L Ă V of dimension dim L “ k ă n, there exists an n ´ k-dimensional
vector space U, a surjective linear map T : V Ñ U and a vector u P U such that
L “ tv P V |T pvq “ uu.
Remark 5.32. If we choose in V and U two bases and we write the linear map by
matrix notation MT v “ u we have the implicit equations of the linear manifold L,
$
’
’
’
’ a11 v1 ` a12 v2 ` ¨ ¨ ¨ ` a1n vn “ u1
’
&
..
’ .
’
’
’
’
%ap1 v1 ` ap2 v2 ` ¨ ¨ ¨ ` apn vn “ up
a1 v1 ` ¨ ¨ ¨ ` an vn “ u0
VL “ tv “ v1 e1 ` ¨ ¨ ¨ ` vn en |f pvq “ 0u “ ker f,
xv, ay “ u0 , where a “ a1 e1 ` ¨ ¨ ¨ ` an en , u0 P R.
$
’
’
’
’ xv, v1 y “ u1
’
&
..
’ .
’
’
’
’
%xv, vp y “ up
The Gram determinant. Distances. 113
where the vectors v1 , . . . vp are linearly independent. The director subspace is given
by
$
’
’
’
’ xv, v1 y “ 0
’
&
..
’ .
’
’
’
’
%xv, vp y “ 0
Remark 5.36. The linear structure implies a very simple but useful fact:
dpv, Uq “ dpv ` w, w ` Uq
for every v, w P V and U Ď V , that is the linear structure implies that the distance
is invariant by translations
}v K } ď }v ´ u} ô
xv K , v K y ď xv K ` v1 ´ u, v K ` v1 ´ uy ô
xv K , v K y ď xv K , v K y ` xv1 ´ u, v1 ´ uy.
b
The second part of the equality, i.e. }v } “ Gpe
K 1 ,...,ek ,vq
Gpe1 ,...,ek q
, follows from the previous
remark.
“ inf dpv ´ v0 , vL q
vL PVL
“ dpv ´ v0 , VL q.
Finally, d
Gpe1 , . . . , ek , v ´ v0 q
dpv, Lq “ dpv ´ v0 , VL q “ ,
Gpe1 , . . . , ek q
where e1 , . . . , ek is a basis in VL .
VL “ tpx, y, z ´ 1, t ´ 2q P R4 |x ` y ` t “ 2, x ´ 2y ` z ` t “ 3u “
tpx, y, z, tq P R4 |x ` y ` t “ 0, x ´ 2y ` z ` t “ 0u,
respectively
VK “ tpx, y, z ´ 2, t ´ 1q P R4 |x ` y ` z ´ t “ 1, x ` y ` z ` t “ 3u “
tpx, y, z, tq P R4 |x ` y ` z ´ t “ 0, x ` y ` z ` t “ 0u.
$
& x`y`t“ 0
By solving the homogenous systems , respectively
% x ´ 2y ` z ` t “ 0
$
& x`y`z´t“0
we obtain that
% x`y`z`t“0
respectively
VK “ spamte3 “ p´1, 1, 0, 0q, e4 “ p´1, 0, 1, 0qu.
Since detre1 |e2 |e3 |e4 s “ 3 ‰ 0 the vectors e1 , e2 , e3 , e4 are linearly independent,
hence VL X VK “ t0u. The distance of v from L is
d c
Gpe1 , e2 , v ´ v0 q 19
dpv, Lq “ dpv ´ v0 , VL q “ “ ,
Gpe1 , e2 q 21
meanwhile d c
Gpe3 , e4 , v ´ v0 q 4
dpv, Kq “ dpv ´ v0 , VK q “ “ .
Gpe3 , e4 q 3
The Gram determinant. Distances. 118
$
’
’ x`y`t“2
’
’
’
’
& x ´ 2y ` z ` t “ 3
It is obvious that K X L ‰ H, since the system is
’
’ x`y`z´t“1
’
’
’
’
% x`y`z`t“3
consistent, having solution p1, 0, 1, 1q, hence we must have
dpL, Kq “ 0.
xv ´ v0 , ny “ 0 .
dpv, Hq “ dpv ´ v0 , VH q.
Let us compute a little now, taking into account the previous observations about
the tangential and normal part:
xv ´ v0 , ny “ xαn ` vH , ny
“ αxn, ny ` xvH , ny
“ α}n}2 ` 0
So, we obtained
|xv ´ v0 , ny|
“ |α|}n} “ }αn}
}n}
The Gram determinant. Distances. 119
that is
|xv ´ v0 , ny|
dpv, Hq “
}n}
In the case that we have an orthonormal basis at hand, the equation of the
hyperplane H is
a1 x1 ` ¨ ¨ ¨ ` ak xk ` b “ 0 ,
|a1 v1 ` ¨ ¨ ¨ ` ak vk ` b|
dpv, Hq “ a .
a21 ` ¨ ¨ ¨ ` a2k
The distance between two linear manifolds
For A and B sets in a metric space, the distance between them is defined as
“ dpv1 ´ v2 , V1 ` V2 q. (5.2)
D1 : x “ x1 ` sd1 ,
D2 : x “ x2 ` td2 ,
M : x “ xM ,
P : x “ xP ` αv1 ` βv2 ,
respectively
H : xx, ny ` b “ 0,
where s, t, α, β, b P R. Recall that two linear manifolds are parallel if the director
space of one of them is included in the director space of the other.
Now we can write down several formulas for distances between linear manifolds.
The Gram determinant. Distances. 121
d
GpxM ´ x1 , d1 q
dpM, D1 q “ ;
Gpd1 q
d
GpxM ´ xP , v1 , v2 q
dpM, P q “ ;
Gpv 1 , v 2 q
d
Gpx1 ´ x2 , d1 , d2 q
dpD1 , D2 q “ if D1 ∦ D2
Gpd1 , d2 q
d
Gpx1 ´ x2 , d1 q
dpD1 , D2 q “ if D1 k D2
Gpd1 , q
|xxM , ny ` b|
dpM, Hq “
}n}
d
Gpx1 ´ xP , d1 , v1 , v2 q
dpD1 , P q “ if D1 ∦ P
Gpd1 , v 1 , v 2 q
But, observe that the distance between these manifolds is actually the distance
between a point M P D and H, hence is more simple to compute from the formula
Problems 122
|xxM ,ny`b|
dpM, Hq “ }n}
, with xM “ u0 . Indeed the equation of H is x ` y ` z ` t “ 1,
thus n “ p1, 1, 1, 1q and b “ ´1, hence
5.6 Problems
Problem 5.6.1. Prove that for the nonzero vectors x, y P R2 , it holds
Problem 5.6.2. Find the angle between the vectors p´2, 4, 3q and p1, ´2, 3q.
Problem 5.6.3. Find the two unit vectors which are orthogonal to both of the
vectors p´2, 3, ´1q and p1, 1, 1q.
for all ai , bi P R , , i “ 1, n.
Problem 5.6.6. Let S be the subspace of the inner product space R3 rXs, the
space of polynomials of degree at most 3, generated by the polynomials 1 ´ x2 and
ş1
2 ´ x ` x2 , where xf, gy “ 0 f pxqgpxqdx. Find a basis for the orthogonal
complement of S.
Problems 123
}u} “ 3 , }u ` v} “ 4 , }u ´ v} “ 6,
find }v}.
Problem 5.6.8. Prove or infirm the following statement: There exists an inner
product on R2 such that the norm induced by this scalar product satisfies
}u ` v}2 ´ }u ´ v}2
xu, vy “ , @u, v P V.
4
Problem 5.6.13. Show that the set of all vectors in Rn which are orthogonal to a
given vector v P Rn is a subspace of Rn . What will its dimension be?
Problem 5.6.15. Let V be an inner product space and let tv1 , . . . , vm u a list of
linearly independent vectors from V. How many orthonormal families te1 , . . . , em uq
can be constructed by using the Gram-Schmidt procedure, such that
spantv1 , . . . , vi u “ spante1 , . . . , ei u, @ i “ 1, m.
Problem 5.6.17. Let V be an inner product space and let U Ď V subspace. Show
that
dim U K “ dim V ´ dim U.
Problem 5.6.20. Find the distance between the following linear manifolds.
Problems 125
d) L “ tpx, y, z, tq P R4 |x ` z ` t “ 1, x ` y ` z “ 2, x ´ y ` t “ 2u, K “
tpx, y, z, tq P R4 |2x ` y ` 2z ` t “ 4u.
Assume now that V is an inner product space. For fixed v P V , the map f : V Ñ F
given by f puq “ xu, vy is a linear functional. The next fundamental theorem shows
that in case V is a Hilbert space, then every linear continuous functional on V is of
this form. Recall that an inner product space is a Hilbert space if is complete, that
is, every Cauchy sequence is convergent. In other words, if the sequence pxn q Ď V
satisfies, the condition:
126
Linear functionals and adjoints 127
f puq “ xu, vy .
Proof. We will present the proof only in the finite dimensional case. We show first
that there is a vector v P V such that f puq “ xu, vy. Let te1 , . . . , en u be an
orthonormal basis of V . One has
Remark 6.3. Note that every linear functional on a finite dimensional Hilbert
space H is continuous. Even more, on every finite dimensional inner product space
H, the inner product defines a norm (metric) such that with the topology induced
by this metric H becomes a Hilbert space.
Linear functionals and adjoints 128
Let us consider another vector space W over F, and an inner product on it, such
that pW, x¨, ¨yq becomes a Hilbert space.
Let T P LpV, W q a continuous operator in the topologies induced by the norms
a a
}v}V “ xv, vyV , respectively }w}W “ xw, wyW , (as a continuous function in
analysis). We define now the adjoint of T , as follows.
Fix w P W . Consider the linear functional on V which maps v in xT pvq, wyW . It
follows that there exists a unique vector T ˚ pwq P V such that
“ yu ` 3zu ` 2xv
Note that in the example above T ˚ is not only a map from R2 to R3 , but also a
linear map.
We shall prove this in general. Let T P LpV, W q, so we want to prove that
T ˚ P LpW, V q.
Let w1 , w1 P W . One has, by definition:
Linear functionals and adjoints 129
which shows that T ˚ pw1 q ` T ˚ pw2 q plays the role of T ˚ pw1 ` w2 q. By the
uniqueness proved before, we have that
“ axv, T ˚pwqy
“ xv, aT ˚pwqy .
This shows that aT ˚ pwq plays the role of T ˚ pawq, and again by the uniqueness we
have that
aT ˚ pwq “ T ˚ pawq .
d) identity I ˚ “ I, if I “ IV , V “ W .
Linear functionals and adjoints 130
For the sake of completeness we prove the above statements. Let v P V and w P W.
aq Let S, T P LpU, W q. Then, xpS ` T qpvq, wy “ xv, pS ` T q˚ pwqy. On the other
hand
xpS`T qpvq, wy “ xSpvq, wy`xT pvq, wy “ xv, S ˚pwqy`xv, T ˚pwqy “ xv, pS ˚`T ˚ qpwqy.
Hence, pS ` T q˚ “ S ˚ ` T ˚ .
bq Let a P F and T P LpU, W q. We have xpaT qpvq, wy “ xv, paT q˚ pwqy. But
xpaT qpvq, wy “ axT pvq, wy “ axv, T ˚ pwq “ xv, aT ˚ pwqy.
Hence, paT q˚ “ aT ˚ pwq.
cq Let T P LpU, W q. Then
xw, T pvqy “ xT pvq, wy “ xv, T ˚pwqy “ xT ˚ pwq, vy “ xw, pT ˚q˚ pvqy.
Hence, pT ˚ q˚ “ T.
dq Let V “ W. We have xv, Ipwqy “ xv, wy “ xIpvq, wy “ xv, I ˚pwqy.
Hence I “ I ˚ .
eq Let T P LpV, W q and S P LpW, Uq. Then for all u P U and v P V it holds:
xT ˚ S ˚ puq, vy “ xS ˚puq, ppT q˚q˚ pvq “ xS ˚ puq, T pvqy “ xu, pS ˚q˚ T pvqy “ xu, ST pvqy “
xST pvq, uy “ xv, pST q˚puqy “ xpST q˚puq, vy.
Hence, T ˚ S ˚ “ pST q˚.
1. ker T ˚ “ pim T qK .
2. im T ˚ “ pker T qK .
3. ker T “ pim T ˚ qK .
4. im T “ pker T ˚ qK .
Linear functionals and adjoints 131
w P ker T ˚ ô T ˚ pwq “ 0
ô xT pvq, wy “ 0 @v P V
ô w P pim T qK ,
The conjugate transpose of a type pm, nq- matrix is an pn, mq matrix obtained by
interchanging the rows and columns and taking the complex conjugate of each
entry. The adjoint of a matrix (which is a linear transform between two finite
dimensional spaces in the appropriate bases) is the conjugate transpose of that
matrix as the next result shows.
pj, kq of MT ˚ the entry is xT ˚ pfk q, ej y, which equals to xfk , T pej qy, which equals to
xT pej q, fk y. In others words, MT ˚ equals to the complex conjugate of MT .
Remark 6.7. We will call a complex square matrix normal if commute with its
conjugate transpose, that is A P Mn pCq is normal iff
AA˚ “ A˚ A,
J
where A˚ is the conjugate transpose of A, that is A˚ “ A . It can easily be
observed that the matrix of a normal operator is a normal matrix.
Example 6.8. On F2 consider the operator which in the canonical basis has the
matrix ¨ ˛
2 ´3
A“˝ ‚.
3 2
This is a normal operator.
T is normal ðñ T ˚ T ´ T T ˚ “ 0
for all w P Epλ0 q. The first term in the inner product lives in Epλ0 q by the
previous statement. Take w “ T ˚ pv0 q ´ λ0 v0 and it follows that T ˚ pv0 q “ λ0 v0 , i.e.
the second assertion in the theorem.
Now follows the last statement. One has T pvq “ λv and T pβq “ βw. By the
previous point T ˚ pwq “ βw, so
(def. of adjoint), which implies λxv, wy “ βxv, wy. Since λ ‰ β, it follows that
xv, wy “ 0.
Proof.
That is T ˚ pwq P U K .
Proof. First suppose that T has a diagonal matrix. The matrix of T ˚ is the
complex transpose, so it is again diagonal. Any two diagonal matrices commutes,
that means that T is normal.
Normal operators 135
To prove the other direction suppose that T is normal. Then, there is a basis
te1 , . . . , en u of V with respect to which the matrix of T is upper triangular, that is
¨ ˛
a a . . . an,n
˚ 1,1 1,2 ‹
˚ ‹
˚ 0 a2,2 . . . a2,n ‹
A“˚ ˚ .. ..
‹
.. ‹ .
˚ . . ¨¨¨ . ‹
˝ ‚
0 0 . . . an,n
and
b
˚
}T pe1 q} “ |a1,1 |2 ` ¨ ¨ ¨ ` |a1,n |2 .
and
b
˚
}T pe2 q} “ |a2,2 |2 ` ¨ ¨ ¨ ` |a2,n |2 .
6.3 Isometries
An operator T P LpV q is called an isometry if
Example 6.14. Let I be the identity map of V (V complex v.s.), and λ P C with
|λ| “ 1. The map λI is an isometry, since }λIpvq} “ }λv} “ |λ|}v} “ }v}.
If T is an isometry it easily follows that T is injective.
Indeed, assume the contrary, that is, there exists u, v P V, u ‰ v such that
T puq “ T pvq. Hence, 0 “ }T puq ´ T pvq} “ }T pu ´ vq} “ }u ´ v}, contradiction with
u ‰ v.
1. T is an isometry.
3. T ˚ T “ I.
6. T ˚ is an isometry.
8. T T ˚ “ I
“ }v}2 .
Remark 6.16. Recall that a real square matrix A is called orthogonal iff
AAJ “ AJ A “ I. A complex square matrix B is called unitary if BB ˚ “ B ˚ B “ I,
J
where B ˚ is the conjugate transpose of B, that is B ˚ “ B . It can easily be
observed that the matrix of an isometry on a real (complex) finite dimensional
inner product space is an orthogonal (unitary) matrix.
The last theorem shows that every isometry is a normal operator. So, the
characterizations of normal operators can be used to give a complete description of
isometries.
Theorem 6.17. Suppose that V is a complex inner product space and T P LpV q.
Isometries 139
}T pvq} “ }v}.
Now we are going to prove the other direction. Suppose T is an isometry. By the
complex spectral theorem there is an orthonormal basis of V consisting of
eigenvectors te1 , . . . , en u.. Let ej , j P t1, . . . , nu be such a vector, associated to an
eigenvalue λj . It follows that
Finally we state the following important theorem concerning on the form of the
matrix of an isometry.
Theorem 6.18. Suppose that V is a real inner product space and T P LpV q. Then
T is an isometry iff there exist an orthonormal basis of V with respect to which T
has a block diagonal matrix where each block on the diagonal matrix is a p1, 1q
matrix containing 1 or ´1, or a p2, 2q matrix of the form
¨ ˛
cos θ ´ sin θ
˝ ‚
sin θ cos θ
Self adjoint operators 140
Remark 6.19. Obviously a self adjoint operator T P LpV q is normal since in this
case holds
T T ˚ “ T ˚ T ˚ “ T ˚ T.
Indeed, for px, yq P F2 one has T px, yq “ p2x ` by, 3x ` 5yq, hence for pu, vq P F2 it
holds
xT px, yq, pu, vqy “ p2x ` byqu ` p3x ` 5yqv “ xpx, yq, p2u ` 3v, bu ` 5vqy.
Remark 6.21. When F “ C the adjoint on LpV q plays a similar role to complex
conjugation on C. A complex number is real iff z “ z. Thus for a self adjoint
operator T the sum T ` T ˚ is analogous to a real number. The analogy is reflected
in some important properties of a self-adjoint operator, beginning with its
eigenvalues.
Remark 6.22. Recall that a complex square matrix A is called hermitian iff
J
A “ A˚ , where A˚ is the conjugate transpose of A, that is A˚ “ A . If A is a
square matrix with real entries, then A is called symmetric iff A “ AJ . It can
easily be observed that matrix of a self adjoint operator on a complex (real) inner
product space is hermitian (symmetric).
λ}v}2 “ xλv, vy
“ xT pvq, vy
“ xv, λvy
“ λ}v}2 .
Problems 142
Theorem 6.24. Let T P LpV q, where V is an inner product space. The following
statements are equivalent.
1. T is self-adjoint.
Proof. Assume that T is self adjoint. Since T is normal there exists exists an
orthonormal basis of V relative to which the matrix of T MT is upper triangular.
But the matrix of T ˚ in this basis is MT ˚ “ MT ˚ , and from T “ T ˚ one has
MT “ MT˚ , hence MT is diagonal, and also the diagonal are formed byy real entries.
Conversely, let MT a diagonal matrix of T , with real entries in some orthonormal
basis. Then MT “ MTJ , hence MT “ MT ˚ or equivalently T “ T ˚ .
6.5 Problems
Problem 6.5.1. Suppose that A is a complex matrix with real eigenvalues which
can be diagonalized by a unitary matrix. Prove that A must be hermitian.
Problem 6.5.2. Prove or give a counter example: the product of any two self
adjoint operators on a finite dimensional inner product space is self adjoint.
Problem 6.5.3. Show that an upper triangular matrix is normal if and only if it
is diagonal.
Problem 6.5.5. Show that if V is a real inner product space, then the set of self
adjoint operators on V is a subspace of LpV q. Show that if V is a complex inner
product space, then the set of self-adjoint operators on V is not a subspace of
LpV q.
Problem 6.5.6. Show that if dim V ě 2 then the set of normal operators on V is
not a subspace of LpV q.
Problem 6.5.7. Let A be a normal matrix. Prove that A is unitary if and only if
all its eigenvalues λ satisfy |λ| “ 1.
Problem 6.5.8. Let v be any unit vector in Cn and put A “ In ´ 2XX ˚ . Prove
that A is both hermitian and unitary. Deduce that A “ A´1 .
im T “ im T ˚ .
and
ker T k “ ker T
im T k “ im T
Problem 6.5.16. Let R2 rXs be the inner product space of polynomials with
degree at most 2, with the scalar product
ż1
xp, qy “ pptqqptqdt.
0
a) Show that the matrix of T with respect to the given basis is hermitian.
(Note that there is no contradiction between these statements because the basis in
the first statement is not orthonormal.)
Qpxq “ X t AX,
where
145
Quadratic forms 146
¨ ˛ ¨ ˛
x1 a11 a12 . . . a1n
˚ ‹ ˚ ‹
˚ ‹ ˚ ‹
˚ x2 ‹ ˚ a12 a22 . . . a2n ‹
X “˚
˚ ..
‹
‹ and A“˚
˚ .. .. ..
‹.
‹
˚ . ‹ ˚ . . . ‹
˝ ‚ ˝ ‚
xn a1n a2n . . . ann
The symmetric matrix A (notice that aij “ aji ) is be called the matrix of the
quadratic form. Being symmetric (and real), A it is the matrix of a self-adjoint
operator with respect to the basis E. This operator, that we call T , is
diagonalizable and there exists a basis B “ tb1 , . . . , bn u formed by eigenvectors
with respect to which T has a diagonal matrix consisting of eigenvalues (also
denoted by T )
T “ diagtλ1 . . . . , λn u.
X “ CX 1
Qpxq “ X J AX
J
“ pCX 1 q A pCX 1 q
“ X 1J C J ACX 1
“ X 1J T X 1
2 2
“ λ1 x1 1 ` ¨ ¨ ¨ ` λn x1 n ,
Quadrics 147
2 2
Qpxq “ λ1 x1 1 ` ¨ ¨ ¨ ` λn x1 n .
7.2 Quadrics
The general equation of a quadric is
From a geometric point of view, quadrics, which are also called quadric surfaces,
are two-dimensional surfaces defined as the locus of zeros of a second degree
Quadrics 148
det R “ 1.
If necessary, i.e., if .det R “ ´1, we must change one of the vectors by its opposite
(for example take R “ r´b1 |b2 |b3 s). This assures that the matrix R defines a
rotation, the new basis being obtained from the original one by this rotation. Let
px, y, zq and px1 , y 1, z 1 q be the coordinates of the same point in the original basis
Quadrics 149
2 2 2
Q “ λ1 x1 ` λ1 y 1 ` λn z 1 ,
and thus, the equation of the quadric reduces to the simpler form
2 2 2
λ1 x1 ` λ1 y 1 ` λn z 1 ` 2a1 14 x1 ` 2a1 24 y 1 ` 2a1 34 z 1 ` a44 “ 0.
To obtain the canonical form of the quadric we still have to perform another
transformation, namely a translation. To complete this step we investigate three
cases: (A) when A has three nonzero eigenvalues, (B) when one eigenvalue is zero
and (C) when two eigenvalues are equal to zero.
(A) For λi ‰ 0 we obtain
λ1 px1 ´ x0 q2 ` λ2 py 1 ´ y0 q2 ` λ3 pz 1 ´ z0 q2 ` a1 44 “ 0
x2 “ x1 ´ x0 ,
y 2 “ y 1 ´ y0 ,
z 2 “ z 1 ´ z0 .
In the new coordinates the equation of the quadric reduces to the canonical form
λ1 x22 ` λ2 y 22 ` λ3 z 22 ` a1 44 “ 0.
7.3 Conics
Studied since the time of ancient greek geometers, conic sections (or just conics)
are obtained, as their name shows, by intersecting a cone with a sectioning plane.
They have played a crucial role in the development of modern science, especially in
astronomy, the motion of earth round the sun taking place on a particular conic
called ellipse. Also, we point out the fact that the circle is a conic section, a special
case of ellipse.
The general equation of a conic is
The following two determinants obtained from the coefficients of the conic play a
crucial role in the classification of conics
ˇ ˇ
ˇ ˇ ˇ ˇ
ˇ a11 a12 a13 ˇ ˇ ˇ
ˇ ˇ ˇ a11 a12 ˇ
ˇ ˇ
∆ “ ˇ a12 a22 a23 ˇ and D2 “ ˇˇ ˇ.
ˇ
ˇ ˇ ˇ a12 a22 ˇ
ˇ ˇ
ˇ a13 a23 a33 ˇ
Notice that the second determinant corresponds to the quadratic form defined by
the first three terms.
x2 y 2
Ellipse pD2 ą 0q whose canonical equation is ` 2 “ 1,
a2 b
Conics 151
x2 y2
Figure 7.1: Ellipse a2
` b2
“1
The reduction of a conic section to its canonical form is very similar with the
procedure that we have presented in the last section when dealing with quadrics.
Again, we must perform a rotation and a translation. We show how the reduction
can be performed by means of an example.
Example 7.1. Find the canonical form of 5x2 ` 4xy ` 8y 2 ´ 32x ´ 56y ` 80 “ 0.
that is
1
x “ ? px1 ´ 2y 1q
5
1
y “ ? p2x1 ` y 1q .
5
To see the translation that we need to perform we rewrite the above equation as
follows
˜ ? ˆ ? ˙2 ¸ ˜ ? ˆ ? ˙2 ¸
8 5 8 5 5 5
9 x12 ´ 2 x1 ` ` 4 y 12 ` 2 y1 `
5 5 5 5
ˆ ? ˙2 ˆ ? ˙2
8 5 5
´9 ´4 ` 80 “ 0.
5 5
Finally, we obtain
ˆ ? ˙2 ˆ ? ˙2
1 8 5 1 5
9 x ´ `4 y ` ´ 30 “ 0.
5 5
? ?
Thus, the translation x2 “ x1 ´ 8 5
5
, y2 “ y1 ` 5
5
reduces the conic to the canonical
form
3 22 2
x ` y 22 “ 1.
10 15
Bibliography
[1] Sheldon Axler: Linear Algebra Done Right, Springer, 2-nd edition, 2004.
153