Further Linear Algebra. Chapter V. Bilinear and Quadratic Forms
Further Linear Algebra. Chapter V. Bilinear and Quadratic Forms
Bilinear
and quadratic forms.
Andrei Yafaev
1 Matrix Representation.
Definition 1.1 Let V be a vector space over k. A bilinear form on V is a
function f : V V k such that
f (u + v, w) = f (u, w) + f (v, w);
f (u, v + w) = f (u, w) + f (u, w).
I.e. f (v, w) is linear in both v and w.
1
P
Recall that if B = {b1 , . . . , bn } is a basis for V and v = xi bi then we
write [v]B for the column vector
x1
[v]B = ... .
xn
Proof. Let
v = x1 b1 + x2 b2 + + xn bn ,
so
x1
[v]B = ... .
xn
Similarly suppose
y1
[w]B = ... .
yn
2
Then
n
!
X
f (v, w) = f xi bi , w
i=1
n
X
= xi f (bi , w)
i=1
n n
!
X X
= xi f bi , yj bj
i=1 j=1
Xn n
X
= xi yj f (bi , bj )
i=1 j=1
n
XX n
= xi yj f (bi , bj )
i=1 j=1
n X
X n
= ai,j xi yj
i=1 j=1
= [v]tB [f ]B [w]B .
Let us give some examples.
Suppose that f : R2 R2 R is given by
x1 x2
f , = 2x1 x2 + 3x1 y2 + x2 y1
y1 y2
Let us write the matrix of f in the standard basis.
3
The matrix
1,1 . . . 1,n
M = ... ..
.
n,1 . . . n,n
is called the transition matrix from B to C. It is always an invertible matrix:
its inverse in the transition matrix from C to B. Recall that for any vector
v V we have
[v]B = M [v]C ,
and for any linear map T : V V we have
[T ]C = M 1 [T ]B M.
Well now describe how bilinear forms behave under change of basis.
[f ]C = M t [f ]B M.
[f ]C = M t AM = M t [f ]B M.
For example, let f be the linear form from the previous example. It is
given by
2 3
1 0
4
in the standard basis. We want to write this matrix in the basis
1 1
,
1 0
5
Theorem 2.2 (Polarization Theorem) If 1 + 1 6= 0 in k then for any
quadratic form q the underlying symmetric bilinear form is unique.
Proof. If u, v V then
q(u + v) = f (u + v, u + v)
= f (u, u) + 2f (u, v) + f (v, v)
= q(u) + q(v) + 2f (u, v).
and
q(x, y) = 2x2 + 2xy = f ((x, y), (x, y))
Let u = (x1 , y1 ), v = (x2 , y2 ) and let us calculate
1 1
(q(u+v)q(u)q(v)) = (2(x1 +x2 )2 +2(x1 +x2 )(y1 +y2 )x21 2x1 y1 2x22 2x1 y2 )
2 2
1
= (4x1 x2 + 2(x1 y2 + x2 y1 )) = f ((x1 , y1 ), (x2 , y2 ))
2
If A = (ai,j ) is a symmetric matrix, then the corresponding form is
X X
f (x, y) = ai,i xi yi + ai,j (xi yj + xj yi )
i i<j
then the symmetric matrix A = (ai,j ) is the matrix representing the under-
lying bilinear form f .
6
3 Orthogonality and diagonalisation
Definition 3.1 Let V be a vector space over k with a symmetric bilinear
form f . We call two vectors v, w V orthogonal if f (v, w) = 0. It is a good
idea to imagine this means that v and w are at right angles to each other.
This is written v w. If S V be a non-empty subset, then the orthogonal
complement of S is defined to be
S = {v V : w S, w v}.
f (v + w, u) = f (v, u) + f (w, u) = 0.
Therefore v + w S .
7
Let U, W be two subspaces of V . The sum of U and W is the subspace
U + W = {u + w : u U, w W }.
We call this a direct sum U W if U W = {0}. This is the same as saying
that ever element of U + W can be written uniquely as u + w with u U
and w W .
Theorem 3.3 (Key Lemma) Let v V and assume that q(v) 6= 0. Then
V = Span{v} {v} .
Proof. For w V , let
f (v, w) f (v, w)
w1 = v, w2 = w v.
f (v, v) f (v, v)
Clearly w = w1 + w2 and w1 Span{v}. Note also that
f (v, w) f (v, w)
f (w2 , v) = f w v, v = f (w, v) f (v, v) = 0.
f (v, v) f (v, v)
Therefore w2 {v} . It follows that Span{v} + {v} = V . To prove that
the sum is direct, suppose that w Span{v} {v} . Then w = v for some
k and we have f (w, v) = 0. Hence f (v, v) = 0. Since q(v) = f (v, v) 6= 0
it follows that = 0 so w = 0.
Proof. [ of the theorem] We use induction on dim(V ) = n. If n = 1 then
the theorem is true, since any 1 1 matrix is diagonal. So suppose the result
holds for vector spaces of dimension less than n = dim(V ).
If f (v, v) = 0 for every v V then using Theorem 5.3 for any basis B we
have [f ]B = [0], which is diagonal. [This is true since
1
f (ei , ej ) = (f (ei + ej , ei + ej ) f (ei , ei ) f (ej , ej )) = 0.]
2
So we can suppose there exists v V such that f (v, v) 6= 0. By the Key
Lemma we have
V = Span{v} {v} .
Since Span{v} is 1-dimensional, it follows that {v} is n 1-dimensional.
Hence by the inductive hypothesis there is an orthonormal basis {b1 , . . . , bn1 }
of {v} .
Now let B = {b1 , . . . , bn1 , v}. This is a basis for V . Any two of the
vectors bi are orthogonal by definition. Furthermore bi {v} , so bi v.
Hence B is an orthogonal basis.
8
4 Examples of Diagonalisation.
Definition 4.1 Two matrices A, B Mn (k) are congruent if there is an
invertible matrix P such that
B = P t AP.
We have shown that if B and C are two bases then for a bilinear form f , the
matrices [f ]B and [f ]C are congruent.
We shall next find out how to calculate the diagonal matrix congruent to
a given symmetric matrix.
There are three kinds of row operation:
multiply row(i) by 6= 0;
multiply column(i) by 6= 0;
9
If E is the corresponging elementary matrix then the double operation
transforms a matrix A into EAE t .
q(xb1 + yb2 ) = x2 y 2 .
Notice that when we have done the first operation, we have multiplied A on
1 0
the left by E2,1 (2) = and when we have done the second, we have
2 1
t 1 2
multiplied on the right by E2,1 (2) =
0 1
We find that
t 1 0
E2,1 (2)AE2,1 (2) =
0 1
Hence in the basis
1 2
,
0 1
1 0
the matrix of the corresponding quadratic form is
0 1
10
Example 4.4 Consider the quadratic form q(x, y)t = 4xy + y 2
0 2 2 1 1 2
2 1 0 2 2 0
1 2 1 0
0 4 0 4
1 0 1 0
0 2 0 1
q(xb1 + yb2 ) = x2 y 2 .
The last step in the previous example transformed the -4 into a -1. In
general, once we have a diagonal matrix we are free to multiply or divide the
diagonal entries by squares:
P = D(1 , . . . , n )
is invertible. Then
Definition 4.3 Two bilinear forms f, f are equivalent if they are the same
up to a change of basis.
Definition 4.4 The rank of a bilinear form f is the rank [f ]B for any basis
B.
Clearly if f and f have different rank then they are not equivalent.
11
5 Canonical forms over C
Definition 5.1 Let q be a quadratic form on vector space V over C, and
suppose there is a basis B of V such that
Ir
[q]B = .
0
Ir
We call the matrix a canonical form of q (over C).
0
Corollary 5.2 Two quadratic forms over C are equivalent iff they have the
same canonical form.
12
Proof. (existence) Let B = {b1 , . . . , bn } be an orthogonal basis. We can
reorder the basis so that
U = Span{b1 , . . . , br }, W = Span{cr +1 , . . . , cn }.
U + W = U W V.
13
The rank of a quadratic form is the rank of the corresponding matrix.
Clearly, in the complex case it is the integer r that appears in the canonical
form.
In the real case, it is r + s.
For a real quadratic form, the signature is the pair (r, s). In this case
q(v) > 0 for all non-zero vectors v.
A real form q is positive definite if its signature is (r, 0), negative
definite if its signature is (0, s). In this case q(v) < 0 for all non-zero
vectors v.
There exists a non-zero vector v such that q(v) = 0 if and only is the
signature is (r, s) with r > 0 and s > 0.
14