0% found this document useful (0 votes)
220 views

Further Linear Algebra. Chapter V. Bilinear and Quadratic Forms

This document discusses bilinear and quadratic forms on vector spaces. It begins by defining bilinear forms and giving examples. It then discusses representing bilinear forms with respect to a basis using matrices. It introduces the change of basis formula for bilinear forms. Finally, it defines symmetric bilinear forms and the associated quadratic forms, proving some basic properties about their relationships.

Uploaded by

Gag Paf
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
220 views

Further Linear Algebra. Chapter V. Bilinear and Quadratic Forms

This document discusses bilinear and quadratic forms on vector spaces. It begins by defining bilinear forms and giving examples. It then discusses representing bilinear forms with respect to a basis using matrices. It introduces the change of basis formula for bilinear forms. Finally, it defines symmetric bilinear forms and the associated quadratic forms, proving some basic properties about their relationships.

Uploaded by

Gag Paf
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 14

Further linear algebra. Chapter V.

Bilinear
and quadratic forms.
Andrei Yafaev

1 Matrix Representation.
Definition 1.1 Let V be a vector space over k. A bilinear form on V is a
function f : V V k such that
f (u + v, w) = f (u, w) + f (v, w);
f (u, v + w) = f (u, w) + f (u, w).
I.e. f (v, w) is linear in both v and w.

An obvious example is the following : take V = R and f : R R R


defined by f (x, y) = xy.
Notice here the difference between linear and bilinear : f (x, y) =
x + y is linear, f (x, y) = xy is bilinear.
More generally f (x, y) = xy is bilinear for any R.
More generally still, given a matrix A Mn (k), the following is a bilinear
form on k n :

v 1 w 1
vi ai,j wj , v = ... , w = ... .
X
f (v, w) = v t Aw. =

i,j vn wn
Well see that in fact all bilenear form are of this form.
 
1 2
Example 1.1 If A = then the corresponding bilinear form is
3 4
      
x1 x2  1 2 x2
f , = x1 y1 = x1 x2 +2x1 y2 +3y1 x2 +4y1 y2 .
y1 y2 3 4 y2

1
P
Recall that if B = {b1 , . . . , bn } is a basis for V and v = xi bi then we
write [v]B for the column vector

x1
[v]B = ... .

xn

Definition 1.2 If f is a bilinear form on V and B = {b1 , . . . , bn } is a basis


for V then we define the matrix of f with respect to B by

f (b1 , b1 ) . . . f (b1 , bn )
[f ]B =
.. ..
. .
f (bn , b1 ) . . . f (bn , bn )

Proposition 1.2 Let B be a basis for a finite dimensional vector space V


over k, dim(V ) = n. Any bilinear form f on V is determined by the matrix
[f ]B . Moreover for v, w V ,

f (v, w) = [v]tB [f ]B [w]B .

Proof. Let
v = x1 b1 + x2 b2 + + xn bn ,
so
x1
[v]B = ... .

xn
Similarly suppose
y1
[w]B = ... .

yn

2
Then
n
!
X
f (v, w) = f xi bi , w
i=1
n
X
= xi f (bi , w)
i=1
n n
!
X X
= xi f bi , yj bj
i=1 j=1
Xn n
X
= xi yj f (bi , bj )
i=1 j=1
n
XX n
= xi yj f (bi , bj )
i=1 j=1
n X
X n
= ai,j xi yj
i=1 j=1

= [v]tB [f ]B [w]B .

Let us give some examples.
Suppose that f : R2 R2 R is given by
   
x1 x2
f , = 2x1 x2 + 3x1 y2 + x2 y1
y1 y2
Let us write the matrix of f in the standard basis.

f (e1 , e1 ) = 2, f (e1 , e2 ) = 3, f (e2 , e1 ) = 1, f (e2 , e2 ) = 0


hence the matrix in the standard basis is
 
2 3
1 0
Now suppose B = {b1 , . . . , bn } and C = {c1 , . . . , cn } are two bases for V .
We may write one basis in terms of the other:
n
X
ci = j,i bj .
j=1

3
The matrix
1,1 . . . 1,n
M = ... ..

.
n,1 . . . n,n
is called the transition matrix from B to C. It is always an invertible matrix:
its inverse in the transition matrix from C to B. Recall that for any vector
v V we have
[v]B = M [v]C ,
and for any linear map T : V V we have

[T ]C = M 1 [T ]B M.

Well now describe how bilinear forms behave under change of basis.

Theorem 1.3 (Change of Basis Formula) Let f be a bilinear form on a


finite dimensional vector space V over k. Let B and C be two bases for V
and let M be the transition matrix from B to C.

[f ]C = M t [f ]B M.

Proof. Let u, v V with [u]B = x, [v]B = y, [u]C = s and [v]C = t.

Let A = (ai,j ) be the matrix representing f with respect to B.


Now x = M s and y = M t so

f (u, v) = (M s)t A(M t)


= (st M t )A(M t)
= st (M t AM )t.

We have f (bi , bj ) = (M t AM )i,j . Hence

[f ]C = M t AM = M t [f ]B M.

For example, let f be the linear form from the previous example. It is
given by  
2 3
1 0

4
in the standard basis. We want to write this matrix in the basis
   
1 1
,
1 0

The transition matrix is :


 
1 1
M=
1 0

its transpose is the same. The matrix of f in the new basis is


 
6 3
5 2

2 Symmetric bilinear forms and quadratic forms.


As before let V be a finite dimensional vector space over a field k.

Definition 2.1 A bilinear form f on V is called symmetric if it satisfies


f (v, w) = f (w, v) for all v, w V .

Definition 2.2 Given a symmetric bilinear form f on V , the associated


quadratic form is the function q(v) = f (v, v).

Notice that q has the property that q(v) = 2 q(v).


For exemple, take the bilinear form f defined by
 
6 0
0 5

The corresponding quadratic form is


 
x
q( ) = 6x2 + 5y 2
y

Proposition 2.1 Let f be a bilinear form on V and let B be a basis for V .


Then f is a symmetric bilinear form if and only if [f ]B is a symmetric matrix
(that means ai,j = aj,i .).

Proof. This is because f (ei , ej ) = f (ej , ei ). 

5
Theorem 2.2 (Polarization Theorem) If 1 + 1 6= 0 in k then for any
quadratic form q the underlying symmetric bilinear form is unique.

Proof. If u, v V then

q(u + v) = f (u + v, u + v)
= f (u, u) + 2f (u, v) + f (v, v)
= q(u) + q(v) + 2f (u, v).

So f (u, v) = 21 (q(u + v) q(u) q(v)). 

Lets look at an example :


Consider  
2 1
A=
1 0
it is a symmetric matrix.
Let f be the corresponding bilinear form. We have

f ((x1 , y1 ), (x2 , y2 )) = 2x1 x2 + x1 y2 + x2 y1

and
q(x, y) = 2x2 + 2xy = f ((x, y), (x, y))
Let u = (x1 , y1 ), v = (x2 , y2 ) and let us calculate
1 1
(q(u+v)q(u)q(v)) = (2(x1 +x2 )2 +2(x1 +x2 )(y1 +y2 )x21 2x1 y1 2x22 2x1 y2 )
2 2
1
= (4x1 x2 + 2(x1 y2 + x2 y1 )) = f ((x1 , y1 ), (x2 , y2 ))
2
If A = (ai,j ) is a symmetric matrix, then the corresponding form is
X X
f (x, y) = ai,i xi yi + ai,j (xi yj + xj yi )
i i<j

and the corresponding quadratic form is


n
X X
q(x) = ai,i x2i + 2 ai,j xi xj
i=1 i<j

then the symmetric matrix A = (ai,j ) is the matrix representing the under-
lying bilinear form f .

6
3 Orthogonality and diagonalisation
Definition 3.1 Let V be a vector space over k with a symmetric bilinear
form f . We call two vectors v, w V orthogonal if f (v, w) = 0. It is a good
idea to imagine this means that v and w are at right angles to each other.
This is written v w. If S V be a non-empty subset, then the orthogonal
complement of S is defined to be

S = {v V : w S, w v}.

Proposition 3.1 S is a subspace of V .

Proof. Let v, w S and k. Then for any u S we have

f (v + w, u) = f (v, u) + f (w, u) = 0.

Therefore v + w S . 

Definition 3.2 A basis B is called an orthogonal basis if any two distinct


basis vectors are orthogonal. Thus B is an orthogonal basis if and only if [f ]B
is diagonal.

Theorem 3.2 (Diagonalisation Theorem) Let f be a symmetric bilinear


form on a finite dimensional vector space V over a field k in which 1 + 1 6= 0.
Then there is an orthogonal basis B for V ; i.e. a basis such that [f ]B is a
diagonal matrix.

Notice that the existence of an orthogonal basis is indeed equivalent to


the matrix being diagonal.
Let B = {v1 , . . . , vn } be an orthogonal basis. By definition f (vi , vj ) = 0 if
i 6= j hence the only possible non-zero values are f (vi , vi ) i.e. on the diagonal.
And of course the converse holds : if the matrix is diagonal, then f (vi , vj ) =
0 of i 6= j.
The quadratic form associated to such a bilinear form is
X
q(x1 , . . . , xn ) = i x2i
i

where i s are elements on the diagonal.

7
Let U, W be two subspaces of V . The sum of U and W is the subspace
U + W = {u + w : u U, w W }.
We call this a direct sum U W if U W = {0}. This is the same as saying
that ever element of U + W can be written uniquely as u + w with u U
and w W .
Theorem 3.3 (Key Lemma) Let v V and assume that q(v) 6= 0. Then
V = Span{v} {v} .
Proof. For w V , let
f (v, w) f (v, w)
w1 = v, w2 = w v.
f (v, v) f (v, v)
Clearly w = w1 + w2 and w1 Span{v}. Note also that
 
f (v, w) f (v, w)
f (w2 , v) = f w v, v = f (w, v) f (v, v) = 0.
f (v, v) f (v, v)
Therefore w2 {v} . It follows that Span{v} + {v} = V . To prove that
the sum is direct, suppose that w Span{v} {v} . Then w = v for some
k and we have f (w, v) = 0. Hence f (v, v) = 0. Since q(v) = f (v, v) 6= 0
it follows that = 0 so w = 0. 
Proof. [ of the theorem] We use induction on dim(V ) = n. If n = 1 then
the theorem is true, since any 1 1 matrix is diagonal. So suppose the result
holds for vector spaces of dimension less than n = dim(V ).
If f (v, v) = 0 for every v V then using Theorem 5.3 for any basis B we
have [f ]B = [0], which is diagonal. [This is true since
1
f (ei , ej ) = (f (ei + ej , ei + ej ) f (ei , ei ) f (ej , ej )) = 0.]
2
So we can suppose there exists v V such that f (v, v) 6= 0. By the Key
Lemma we have
V = Span{v} {v} .
Since Span{v} is 1-dimensional, it follows that {v} is n 1-dimensional.
Hence by the inductive hypothesis there is an orthonormal basis {b1 , . . . , bn1 }
of {v} .
Now let B = {b1 , . . . , bn1 , v}. This is a basis for V . Any two of the
vectors bi are orthogonal by definition. Furthermore bi {v} , so bi v.
Hence B is an orthogonal basis. 

8
4 Examples of Diagonalisation.
Definition 4.1 Two matrices A, B Mn (k) are congruent if there is an
invertible matrix P such that

B = P t AP.

We have shown that if B and C are two bases then for a bilinear form f , the
matrices [f ]B and [f ]C are congruent.

Theorem 4.1 Let A Mn (k) be symmetric, where k is a field in which


1 + 1 6= 0, then A is congruent to a diagonal matrix.

Proof. This is just the matrix version of the previous theorem. 

We shall next find out how to calculate the diagonal matrix congruent to
a given symmetric matrix.
There are three kinds of row operation:

swap rows i and j;

multiply row(i) by 6= 0;

add row(i) to row(j).

To each row operation there is a corresponding elementary matrix E; the


matrix E is the result of doing the row operation to In . The row operation
transforms a matrix A into EA.
We may also define three corresponding column operations:

swap columns i and j;

multiply column(i) by 6= 0;

add column(i) to column(j).

Doing a column operation to A is the same a doing the corresponding row


operation to At . We therefore obtain (EAt )t = AE t .

Definition 4.2 By a double operation we shall mean a row operation fol-


lowed by the corresponding column operation.

9
If E is the corresponging elementary matrix then the double operation
transforms a matrix A into EAE t .

Lemma 4.2 If we do a double operation to A then we obtain a matrix con-


gruent to A.

Proof. EAE t is congruent to A. 

Recall that a symmetric bilinear forms are represented by symmetric ma-


trices. If we change the basis then we will obtain a congruent matrix. Weve
seen that if we do a double operation to matrix A then we obtain a congruent
matrix. This corresponds to the same quadratic form with respect to a dif-
ferent basis. We can always do a sequence of double operations to transform
any symmetric matrix into a diagonal matrix.

Example 4.3 Consider the quadratic form q(x, y)t = x2 + 4xy + 3y 2


     
1 2 1 2 1 0
A= .
2 3 0 1 0 1

This shows that there is a basis B = {b1 , b2 } such that

q(xb1 + yb2 ) = x2 y 2 .

Notice that when we have done the first operation, we have multiplied A on
1 0
the left by E2,1 (2) = and when we have done the second, we have
2 1  
t 1 2
multiplied on the right by E2,1 (2) =
0 1
We find that  
t 1 0
E2,1 (2)AE2,1 (2) =
0 1
Hence in the basis    
1 2
,
0 1
 
1 0
the matrix of the corresponding quadratic form is
0 1

10
Example 4.4 Consider the quadratic form q(x, y)t = 4xy + y 2
     
0 2 2 1 1 2

2 1 0 2 2 0
   
1 2 1 0

0 4 0 4
   
1 0 1 0

0 2 0 1

This shows that there is a basis {b1 , b2 } such that

q(xb1 + yb2 ) = x2 y 2 .

The last step in the previous example transformed the -4 into a -1. In
general, once we have a diagonal matrix we are free to multiply or divide the
diagonal entries by squares:

Lemma 4.5 For 1 , . . . , n k = k \ {0} and 1 , . . . , n k

D(1 , . . . , n ) is congruent to D(21 1 , . . . 2n n ).

Proof. Since 1 , . . . , n k \ {0} then 1 n 6= 0. So

P = D(1 , . . . , n )

is invertible. Then

P t D(1 , . . . , n )P = D(1 , . . . , n )D(1 , . . . , n )D(1 , . . . , n )


= D(21 1 , . . . , 2n n ).

Definition 4.3 Two bilinear forms f, f are equivalent if they are the same
up to a change of basis.

Definition 4.4 The rank of a bilinear form f is the rank [f ]B for any basis
B.

Clearly if f and f have different rank then they are not equivalent.

11
5 Canonical forms over C
Definition 5.1 Let q be a quadratic form on vector space V over C, and
suppose there is a basis B of V such that
 
Ir
[q]B = .
0
 
Ir
We call the matrix a canonical form of q (over C).
0

Theorem 5.1 (Canonical forms over C) Let V be a finite dimensional


vector space over C and let q be a quadratic form on V . Then q has exactly
one canonical form.

Proof. (Existence) We first choose an orthogonal basis B = {b1 , . . . , bn }.


After reordering the basis we may assume that q(b1 ), . . . , q(br ) 6= 0 and
q(br+1 ), . . . , q(bn ) = 0. p
Since every complex number has a square root in
C, we may divide bi by q(bi ) if i r.
(Uniqueness) Change of basis does not change the rank. 

Corollary 5.2 Two quadratic forms over C are equivalent iff they have the
same canonical form.

6 Canonical forms over R


Definition 6.1 Let q be a quadratic form on vector space V over R, and
suppose there is a basis B of V such that

Ir
[q]B = Is .
0

Ir
We call the matrix Is a canonical form of q (over R).
0

Theorem 6.1 (Sylvesters Law of Inertia) Let V be a finite dimensional


vector space over R and let q be a quadratic form on V . Then q has exactly
one (real) canonical form.

12
Proof. (existence) Let B = {b1 , . . . , bn } be an orthogonal basis. We can
reorder the basis so that

q(b1 ), . . . , q(br ) > 0, q(br+1 ), . . . , q(br+s ) < 0, q(br+s+1 ), . . . , q(bn ) = 0.

Then define a new basis by


1
(
bi i r + s,
|q(bi )|
ci =
bi i > r + s.

The matrix of q with respect to C is a canonical form.


(uniqueness) Suppose we have two bases B and C with

Ir Ir
[q]B = Is , [q]C = Is .
0 0

By comparing the ranks we know that r + s = r + s . Its therefore sufficient


to prove that r = r . Define two subspaces of V by

U = Span{b1 , . . . , br }, W = Span{cr +1 , . . . , cn }.

If u is a non-zero vector of U then we have u = x1 b1 + . . . + xr br . Hence

q(u) = x21 + . . . + x2r > 0.

Similarly if w W then w = yr +1 cr +1 + . . . + yn cn , and

q(w) = yr2 +1 . . . yr2 +s 0.

It follows that U W = {0}. Therefore

U + W = U W V.

From this we have


dim U + dim W dim V.
Hence
r + (n r ) n.
This implies r r . A similar argument (consider U = Span{c1 , . . . , cr } and
W = Span{br+1 , . . . , bn }) shows that r r, so we have r = r . 

13
The rank of a quadratic form is the rank of the corresponding matrix.
Clearly, in the complex case it is the integer r that appears in the canonical
form.
In the real case, it is r + s.
For a real quadratic form, the signature is the pair (r, s). In this case
q(v) > 0 for all non-zero vectors v.
A real form q is positive definite if its signature is (r, 0), negative
definite if its signature is (0, s). In this case q(v) < 0 for all non-zero
vectors v.
There exists a non-zero vector v such that q(v) = 0 if and only is the
signature is (r, s) with r > 0 and s > 0.

14

You might also like