A Generalization of Cramers Rule and Applications
A Generalization of Cramers Rule and Applications
net/publication/237323822
CITATION READS
1 735
2 authors, including:
Hugo Leiva
Yachay Tech-University of the Andes (Venezuela)
195 PUBLICATIONS 1,852 CITATIONS
SEE PROFILE
All content following this page was uploaded by Hugo Leiva on 16 April 2015.
Hugo Leiva
Abstract
In this paper we find a formula for the solutions of the following linear equation
Ax = b, x ∈ IRm , b ∈ IRn , m ≥ n,
where A = (aj,i )n×m is a n × m real matrix. We prove the following statement: For all
b ∈ IRn the system is solvable if, and only if, the set of vectors {l1 , l2 , . . . , ln } formed by the
rows of the matrix A is lineally independent in IRm . Moreover, one solution for this equation
is given by the following formula
Resumen
En este artículo se encuentra una formula para las soluciones de la siguiente ecuación lineal
Ax = b, x ∈ IRm , b ∈ IRn , m ≥ n,
donde A = (aj,i )n×m es una matriz real n × m. Probaremos el siguiente resultado: Para todo
b ∈ IRn el sistema soluble si, y sólo si, el conjunto de vectores {l1 , l2 , . . . , ln } formado por las
81
82 Hugo Leiva
filas de la matriz A es linealmente independiente en IRm . Más aún, una solución para esta
ecuación viene dada por la siguiente fórmula
1 Introduction
In this paper we find a formula for the solutions of the following linear equation
or
a1,1 x1 + a1,2 x2 + · · · + a1,m xm = b1
a2,1 x1 + a2,2 x2 + · · · + a2,m xm = b2
.. .. .. .. .. (2)
. . . . .
an,1 x1 + an,2 x2 + · · · + an,m xm = bn
where < ·, · > denotes the innerproduct in IRm and A is n × m real matrix. Usually, one can
apply Gauss Elimination Method to find some solutions of this system, this method is a systematic
procedure for solving systems like (1); it is based on the idea of reducing the augmented matrix
a1,1 a1,2 · · · a1,m b1
a2,1 a2,2 · · · a2,m b2
.. .. .. .. .. , (4)
. . . . .
an,1 an,2 · · · an,m bn
to the form that is simple enough such that the system of equations can be solved by inspec-
tion. But, to my knowledge, in general there is not formula for the solutions of (1) in terms of
determinants if m 6= n.
When m = n and det(A) 6= 0 the system (1) admits only one solution given by x = A−1 b,
and from here one can deduce the well known Cramer Rule which says:
Theorem 1.1 (Cramer Rule 1704-1752) If A is n × n matrix with det(A) 6= 0, then the solution
of the system (1) is given by the formula:
det((A)i )
xi = , i = 1, 2, 3, . . . , n, (5)
det(A)
where (A)i is the matrix obtained by replacing the entries in the ith column of A by the entries
in the matrix
b1
b2
.. .
.
bn
A simple and interested generalization of Cramer Rule was done by Prof. Dr. Sylvan Burgstahler
([2]) from University of Minnesota, Duluth, where he taught for 20 years. This result is given by
the following Theorem:
Theorem 1.3 For all b ∈ IRn the system(1) is solvable if, and only if,
det(AA∗ ) 6= 0. (8)
Moreover, one solution for this equation is given by the following formula
Also, this solution coincides with the Cramer formula when n = m. In fact, this formula are
given as follows:
n
X det((AA∗ )j )
xi = aj,i , i = 1, 2, 3, . . . , m, (10)
det(AA∗ )
j=1
where (AA∗ )j is the matrix obtained by replacing the entries in the jth column of AA∗ by the
entries in the matrix
b1
b2
.. .
.
bn
In addition, this solution has minimum norm. i.e.,
Theorem 1.4 The solution of (1)- (3) given by (9) can be written as follows:
Theorem 1.5 The system (1) is solvable for each b ∈ IRn if, and only if, the set of vectors
{l1 , l2 , . . . , ln } formed by the rows of the matrix A is lineally independent in IRm .
Moreover, a solution for the system (1) is given by the following formula:
υ1i υ2i < l2 , υ1 >
xi = b1 + b2 − c1 (13)
kυ1 k2 kυ2 k2 kυ1 k2
n−1
!
υni X < ln , υi >
+ +··· + bn − ci , i = 1, 2, . . . , m, (14)
kυn k2 kυi k2
i=1
where the set of vectors {υ1 , υ2 , . . . , υn } is obtain by the Gram-Schmidt process and the numbers
c1 , c2 , . . . , cn are given by
c1 = b1 (15)
< l2 , υ1 >
c2 = b2 − c1
kυ1 k2
< l3 , υ1 > < l3 , υ2 >
c3 = b3 − 2
c1 − c2
kυ1 k kυ2 k2
.. .. .. .. ..
. . . . .
n−1
X < ln , υi >
cn = bn − ci ,
kυi k2
i=1
Finally, we apply these results to find a solutions of the following generalized linear differential
equation
In this section we shall prove Theorems 1.3, 1.4, 1.5 and more. To this end, we shall denote by
√
< x, y > the Euclidian innerproduct in IRk and the associated norm by k xk = < x, x >. Also,
we shall use some ideas from [3] and the following result from [1], pp 55.
Lemma 2.1 Let W and Z be Hilbert space, G ∈ L(W, Z) and G∗ ∈ L(Z, W ) the adjoint opera-
tor, then the following statements holds,
Proof of Theorem 1.3. The matrix A may also viewed as a linear operator A : IRm → IRn ;
therefore A ∈ L(IRm , IRn ) and its adjoint operator A∗ is the transpose of A and A∗ : IRn → IRm .
Then, system (1) is solvable for all b ∈ IRn if, and only if, the operator A is surjective. Hence,
from the Lemma 2.1 there exists γ > 0 such that
Therefore,
< AA∗ z, z >≥ γ 2 kzk2IRn , z ∈ IRn .
This implies that AA∗ is one to one. Since AA∗ is a n × n matrix, then det(AA∗ ) 6= 0.
Suppose now that det(AA∗ ) 6= 0. Then (AA∗ )−1 exists and given b ∈ IRn we can see that
x = A∗ (AA∗ )−1 b is a solution of Az = b.
(AA∗ )w = b,
where (AA∗ )i is the matrix obtained by replacing the entries in the ith column of AA∗ by the
entries in the matrix
b1
b2
.. .
.
bn
Then, the solution x = A∗ (AA∗ )−1 b of (1) can be written as follows
det((AA∗ )1 ) det((AA∗ )j )
P
n
det(AA∗ ) j=1 aj,1 det(AA∗ )
a1,1 a2,1 · · · an,1
det((AA∗ )2 ) det((AA∗ )j )
P
a1,2 a2,2 · · · n
an,2
j=1 aj,2 det(AA∗ )
x= .
det(AA∗ )
=
.
.. .. ..
.. ..
.. . . . .
.
a1,m a2,m · · · an,m det((AA∗ )n ) det((AA∗ )j )
P
n
det(AA∗ ) j=1 aj,m det(AA∗ )
Now, we shall see that this solution has minimum norm. In fact, consider w in IRm such that
Aw = b and
< x, w − x >=< A∗ (AA∗ )−1 b, w − x =< (AA∗ )−1 b, Aw − Ax >=< (AA∗ )−1 b, b − b >= 0.
where
det((AA∗ )j )
zj = , j = 1, 2, . . . , n,
del(AA∗ )
is the only solution of the system
AA∗ z = b,
Proof of Theorem 1.5. Suppose the system is solvable for all b ∈ IRn . Now, assume the
existence of real numbers ci , i = 1, 2, . . . , n such that
c1 l1 + c2 l2 + c3 l3 + · · · + cn ln = 0.
In other words,
< li , x >= ci , i = 1, 2, . . . , n.
Hence,
< ci li , x >= c2i , i = 1, 2, . . . , n.
A Generalization of Cramer’s Rule and Applications to Generalized Linear Differential Equations. 89
So,
< c1 l1 + c2 l2 + c3 l3 + · · · + cn ln , x >= c21 + c22 + c23 + · · · + c2n = 0.
Now, suppose that the set {l1 , l2 , . . . , ln } is linearly independent in IRm . Using the Gram-
Schmidt process we can find a set {υ1 , υ2 , . . . , υn } of orthogonal vectors in IRm given by the
formula:
υ1 = l1 (17)
< l2 , υ1 >
υ2 = l2 − υ1
kυ1 k2
< l3 , υ1 > < l3 , υ2 >
υ3 = l3 − υ1 − υ2
kυ1 k2 kυ2 k2
.. .. .. .. ..
. . . . .
n−1
X < ln , υi >
υn = ln − υi
kυi k2
i=1
where
c1 = b1 (19)
< l2 , υ1 >
c2 = b2 − c1
kυ1 k2
< l3 , υ1 > < l3 , υ2 >
c3 = b3 − 2
c1 − c2
kυ1 k kυ2 k2
.. .. .. .. ..
. . . . .
n−1
X < ln , υi >
cn = bn − ci
kυi k2
i=1
In this section we shall consider some particular cases and examples to illustrate the results of
this work.
In this case n = 1 and A = [a1,1 , a1,2 , · · · , a1,m ]. Then, if we define the column vector
a1,1
a1,2
l1 = . ,
. .
a1,m
a1,1
i a1,2
AA∗ = a1,1 a1,2 ... a1,m . = kl1 k2 .
h
..
a1,m
Then, (AA∗ )−1 b = bkl1 k−2 and
a1,1 bkl1 k−2
a1,2 bkl1 k−2
x = A∗ (AA∗ )−1 b = .. .
.
a1,m bkl1 k−2
A Generalization of Cramer’s Rule and Applications to Generalized Linear Differential Equations. 91
then det(AA∗ ) = kl1 k2 kl2 k2 − | < l1 , l2 > |2 = kl1 k2 kl2 k2 = 6 and x1 = 56 , x2 = 16 , and x3 = −2
6 .
kl1 k2
0 0 ··· 0
0 2
kl2 k 0 · · · 0
∗
.. .
. .
. .
. ..
AA = . . . . . ,
.. .. .. .. ..
. . . . .
0 0 0 · · · kln k2
and the solution of the system (1) is very simple and given by:
n
X
xi = aj,i bj klj k−2 , i = 1, 2, . . . , m. (29)
j=1
Now, we shall apply the formula (29) to find solution of the following system
−x1 − x2 + x3 + x4 = 1
−x1 + x2 − x3 + x4 = 1 (30)
x1 − x2 − x3 + x4 = 1
1 1 1
Then, {l1 , l2 , L3 } is an orthogonal set in IR3 and the solution of this system is given by: x1 = −1
4 ,
−1 −1
x2 = 4 , x3 = 4 and x4 = 34 .
A Generalization of Cramer’s Rule and Applications to Generalized Linear Differential Equations. 93
The Theorems 1.3, 1.4 and 1.5 give a formula for one solution of the system (1) which has
minimum norma. But, it is no the only way allowing to build solutions of this equation. Next,
we shall present a variational method to obtain solutions of (1) as a minimum of a quadratic
functional : IRn → IR,
1 ∗ 2
(ξ) = kA ξk − < b, ξ >, ∀ξ ∈ IRn . (31)
2
Proposition 3.1 For a given b ∈ IRn the equation (1) has a solution x ∈ IRm if, and only if,
It is easy to see that (32) is in fact an optimality condition for the critical points of the quadratic
functional define above.
Lemma 3.1 Suppose the quadratic functional has a minimizer ξb ∈ IRn . Then,
xb = A∗ ξb , (33)
is a solution of (1).
Remark 3.1 Under the condition of Theorem 1.3, the solution given by the formulas (33) and (
9) coincide.
Theorem 3.1 The system ( 8) is solvable if, and only if, the quadratic functional define by (
31) has a minimum for all b ∈ IRn
94 Hugo Leiva
Proof . Suppose ( 8) is solvable. Then, the matrix A viewed as an operator from IRm to IRn is
surjective. Hence, from Lemma 2.1 there exists γ > 0 such that
Then,
γ2
(ξ) ≥ kξk2 − kbkkξk, ξ ∈ IRn .
2
Therefore,
lı́m (ξ) = ∞.
kξk→∞
Consequently, is coercive and the existence of a minimum is ensured.
Now, we shall consider an example where Theorems 1.3, 1.4 and 1.5 can not be applied, but
proposition 3.1 does.
AA∗ ξ = b.
i.e.,
3ξ1 + 6ξ2 = 1
6ξ1 + 12ξ2 = 2
So, there are infinitely many critical points given by
1
− 2a
ξ= 3 , a ∈ IR.
a
Hence, a solution of the system is given by
1
1 2 1
3
− 2a
x = A∗ ξ = 1 2 3 = 1
3
a 1
1 2 3
A Generalization of Cramer’s Rule and Applications to Generalized Linear Differential Equations. 95
The case m < n is undetermined since the equation Ax = b has solution only when b ∈ Rang(A).
But, nevertheless we can produce the following Theorems:
Theorem 4.1 Suppose m < n and A : IRm → IRn is one to one. Then, for all b ∈ Rang(A) the
equation
where (A∗ A)i is the matrix obtained by replacing the entries in the ith column of AA∗ by the
entries in the matrix Pn
aj,1 bj
Pj=1
n
j=1 aj,2 bj
..
Pn .
j=1 aj,m bj
Proof . Since A is one to one, from Lemma 2.1 we have that A∗ : IRn → IRm is surjective, and
consequently det(A∗ A) 6= 0. Therefore, (A∗ A)−1 exists and
is the only solution of (34). In fact, if x ∈ IRm is the only solution of (34), then A∗ Ax = A∗ b and
(A∗ A)−1 A∗ Ax = (A∗ A)−1 A∗ b. So, x = (A∗ A)−1 A∗ b. The remainder of the proof follows in the
same way as in Theorem 1.3
Theorem 4.2 If the set of vectors {l1 , l2 , . . . , ln } formed by the columns of the matrix A is an
orthogonal set in IRm , then
kl1 k2
0 ··· 0
0 2
kl2 k · · · 0
A∗ A = .. .. .. .. .
. . . .
0 0 ··· kln k2
96 Hugo Leiva
and the solution x = (A∗ A)−1 A∗ b of the system (34) takes the following simple form
1 Pn
a b
kl1 k2 Pj=1 j,1 j
1 n
kl2 k2 j=1 aj,2 bj
x= .. .
.
1 Pn
klm k 2 j=1 aj,m bj
Let A and B be n × m matrixes with m ≥ n such that det(AA∗ ) 6= 0 and the eigenvalues of
BA∗ (AA∗ )−1 are not zero. Under these conditions we consider the following generalized linear
system of differential equations with implicit derivative
Before study the non homogeneous equation (38) we shall look for the solution of the homogeneous
equation
Then, if we putt S = A∗ (AA∗ )−1 we get the following algebraic equations for λ and b:
det(λBS − I) = 0. (41)
Lemma 5.1 If α is an eigenvalue of the matrix BS = BA∗ (AA∗ )−1 with corresponding eigen-
vector b ∈ IRn , then x(t) = eλt ξ, with λ = α−1 and ξ = Sb, is a solution of (39).
1
Corollary 5.1 If BS possess n linearly independent eigenvectors b1 , b2 , . . . , bn and λj be the real
1 1 1
eigenvalue that corresponds to bj (The numbers λ1 , λ2 , . . . , λn need not all be distinct). Then, for
all cj ∈ IR, j = 1, 2, . . . , n
n
X
x(t) = cj eλj t ξj , ξj = Sbj , (42)
j=1
Corollary 5.2 Under the conditions of the foregoing Theorem, if b1 , b2 , . . . , bn are orthogonal,
then a general solution of (38) is given by
n
X Z t Xn
x(t) = cj eλj t ξj + eλj (t−s) < bj , f (s) > ξj ds, ξj = Sbj , (45)
j=1 0 j=1
Proof of Theorem 5.1.. Clearly that a solution x(t) of (38) has the form x(t) = xh (t) + xp (t),
where xh (t) is a solution of the homogeneous equation (39) and xp (t) is a particular solution of
the non homogeneous equation (38). So, we look for a particular solution of (38) by the variation
constant method; that is to say, we need to find functions cj (t), j = 1, 2, . . . , n such that
n
X
x(t) = cj (t)eλj t ξj ,
j=1
Referencias
[1] R.F. CURTAIN and A.J. PRITCHARD, “Infinite Dimensional Linear Systems”, Lecture
Notes in Control and Information Sciences, Vol. 8. Springer Verlag, Berlin (1978).
[3] E. ITURRIAGA and H. LEIVA, “A Necessary and Sufficient Conditions for the Controlla-
bility of Linear System in Hilbert Spaces and Applications” IMA Journal Mathematical and
Information, PP.1-12,doi:10.1093/imamci/dnm017 (2007).
HUGO LEIVA
Departamento de Matemáticas, Facultad de Ciencias,
Universidad de Los Andes
Mérida 5101, Venezuela
e-mail: [email protected]