LU-factorization and Positive Definite Matrices: Tom Lyche
LU-factorization and Positive Definite Matrices: Tom Lyche
Definite Matrices
Tom Lyche
University of Oslo
Norway
1 2 3
i 1 2 3
A11 A12 h
(i) = 4 5 6 , (ii) a.1 , a.2 , a.3 =
4 ,
5 6
A21 A22
7 8 9 7 8 9
aT1. 1 2 3
h i 1 2 3
(iii) T =
2. 4
a ,
5 6 (iv) A11 , A12 =
4 5 6 .
aT3. 7 8 9 7 8 9
One way to see that this is correct is to partition A into columns and
x into rows.
where A, A11 and A22 are square matrices. Then A is nonsingular if and only if both
A11 and A22 are nonsingular. In that case
−1 −1 −1
A 11 −A 11 A12 A 22
A−1 = (1)
0 A−1
22
The first equation implies that A11 is invertible, this in turn implies that
B 21 = 0 in the second equation, and then the third equation simplifies to
B 22 A22 = I. We conclude that also A22 is invertible.
Proof. Exercise.
Proof. x
But then both matrices must be diagonal with ones on the diagonal.
−1
We conclude that L2 L1 = I = U 1 U −1
2 which means that L1 = L2 and
U 1 = U 2.
where D k , Nk , Wk ∈ Rn−k,n−k .
Ak = Lk U k (3)
B k = Lk Vk (4)
C k = Mk U k (5)
D k = Mk Vk + Nk Wk (6)
Since Lk is unit lower triangular and U k is upper triangular we see that (3)
gives the LU -factorization of Ak .
xT Ax = 0 ⇒ x = 0. (7)
0 −1 2
T T T
2 T
x Ax = x B Bx = (Bx) (Bx) =
Bx
2 . (9)
Since the Euclidian norm k k2 of a vector is nonnegative this shows that A is positive
semidefinite and that A is positive definite if and only if B has linearly independent
columns.
Note that A and B have the same null-space and the same rank.
It is shown in advanced calculus texts that under suitable conditions on the domain Ω the
matrix H(t) is symmetric for each t ∈ Ω. Moreover if ∇F (t∗ ) = 0 and H(t∗ ) is
positive definite then t∗ is a local minimum for F . This can be shown using second-order
Taylor approximation of F . Moreover, t∗ is a local maximum if ∇F (t∗ ) = 0 and H(t∗ )
is negative definite.
A1 : If a diagonal entry aii ≤ 0 for some i then eTi Aei = aii ≤ 0 and
A is not positive definite.
A2 : if the absolute value of the largest entry of A is not (only) on the
diagonal then A is not positive definite.
To show this suppose aij ≥ aii and aij ≥ ajj for some i 6= j. Since A
is symmetric we obtain (ei − ej )T A(ei − ej ) = aii + ajj − 2aij ≤ 0
which implies that xT Ax ≤ 0 for some x 6= 0.
A = RT R, RT := LD 1/2 , (10)
1/2 1/2
where D 1/2 = diag(d11 , . . . , dnn ). Since L and D 1/2 are nonsingular it
follows that R is nonsingular and A is positive definite by Example 4.
Proof. x
Pn
3. For any x ∈ Rn we have Ax = j=1 λj cj uj , where λ1 , . . . , λn are
the eigenvalues of A.
Pn
4. Furthermore hAx, xi = j=1 λj c2j .
5. cond2 (A) := kAk2 kA−1 k2 = λλmax
min
, where λmax and λmin are the
largest and smallest eigenvalue of A.
n min(j,k)
X X
akj = rik rij = rik rij , j, k = 1, . . . , n. (12)
i=1 i=1
k−1
2 1/2
X
rkk = akk − rik ,
i=1
(13)
k−1
X
rkj = akj − rik rij /rkk j = k + 1, . . . , n.
i=1
for k = 1, 2, . . . , n
s = R(1:k−1, k); R(k, k) = (A(k, k) − sT ∗ s)1/2 ;
T
R(k, k+1:n) = A(k, k+1:n) − s ∗ R(1:k−1, k+1:n) /R(k, k);
end
Proof. We show that if R has upper bandwidth d in its first k − 1 rows then row k also
has upper bandwidth d. The proof then follows by induction on k . Now, if j > k + d, then
akj = 0, and if R has upper bandwidth d in its first k − 1 rows then for i > k + d we
have rji = 0 for j = 1, . . . , k − 1. From(13) it follows that rki = 0 for i > k + d.
for k = 1, 2, . . . , n
s = R(1:k−1, k); R(k, k) = (A(k, k) − sT ∗ s)1/2 ;
T
R(k, k+1:n) = A(k, k+1:n) − s ∗ R(1:k−1, k+1:n) /R(k, k);
end
Banded version:
for k = 1, 2, . . . , n
km = max(1, k − d); kp = min(n, k + d);
s = R(km:k−1, k); R(k, k) = sqrt(A(k, k) − sT ∗ s);
T
R(k, k+1:kp) = A(k, k+1:kp) − s ∗ R(km:k−1, k+1:kp) /R(k, k);
end