1 s2.0 0024379578900757 Main
1 s2.0 0024379578900757 Main
Roush
Alabama State University
Montgomery, Alabamu 36101
ABSTRACT
1. INTRODUCTION
We give algorithms (i) for finding all g-inverses of a Boolean matrix, (ii)
for finding all minimum-norm g-inverses of a Boolean matrix, (iii) for finding
all least-squares g-inverses of a Boolean matrix, and (iv) for finding all
Vagner inverses of a Boolean matrix. We give a criterion for a Boolean
matrix to be regular, and criteria for Boolean matrices to have minimum-
norm g-inverses, least-squares g-inverses, and Moore-Penrose inverses. We
count the number of Boolean matrices of specified rank and dimension
having one of these last types of inverses. These inverses have applications in
network and switching theory and information theory [3].
2. PRELIMINARIES
TABLE 1
DEFINITIONS OF INVERSES
g-inverse AGA=A
Minimum-norm g-inverse AGA=A, (GA)T=GA
Least-squares g-inverse AGA = A, (AG)T= AC
Vagner inverse AGA=A, GAG= G
Moore-Penrose inverse AGA = A, GAG = G,
IAGJT=AG, (GA)*=GA
The second and third inverses were named by P. Rao and B. Rao [12].
The vagner inverse was introduced by Vagner [15] and is of importance in
general semigroup theory. Other authors have called it semi-inverse [8],
Moore-Penrose inverse [4], and reflexive g-inverse [lo, 11, 121.
4. KNOWN-RESULTS
Only since 1971 has the field of study of g-inverses and Vagner inverses
been investigated systematically. The concept of inverse of matrix was first
introduced by Wedderburn [16] as early as 1934: G is said to be an inverse
of A if AC = GA = Z, where Z is an identity matrix. In 1952, Lute [S] showed
that A possesses a two-sided inverse if and only if A is an orthogonal matrix
in the sense that AA r= I, and that in this case A r is the two-sided inverse.
In 1963, Rutherford [13] showed that if A possesses a one-sided inverse, that
inverse is also a two-sided inverse. Furthermore such an inverse, if it exists, is
unique and is A r. It is easy to see that the permutation matrices are the only
matrices (over p) which have inverses.
250 KI HANG KIM AND FRED W. ROUSH
In 1971, Plemmons [8] gave the following method for finding the Vagner
inverses of a matrix: find idempotent E in the C-class containing A, and
solve the equation XA = E. In 1973, unaware of the earlier work of Plem-
mons [a], Kim [4] showed that if A is nonsingular, then the Vagner inverse of
A is the adjoint A* of A, defined by A* = [Ai J, where Aii is the cofactor of
uji in
Here the summation is taken in the Boolean algebra over all permutations u
of 1,.,., n. Thus ]A] is either 0 or 1.
In 1973, P. Rao and B. Rao [12] studied many types of inverses by
writing regular matrices in the form
Pi A,’ 0
[0 OQ
1
5. ALGORITHMS
EXAMPLE1. Let A be
1 1
1 1 1
0 11.
1 0 0
(1) Find b(A). A minimum-norm g-inverse will exist if and only if the
sets of l’s in the members of b(A) are disjoint, i.e.. for v,tu E b(A), u# w,
uj = 1 imphes wj =O.
(2) Let E be the matrix with eii = 1 if and only if for some u E b(A),
t$ = vi = 1.
(3) Any solution G of GA = E will be minimum-norm g-inverse of A. To
find such solutions. for each row of E. determine which linear combinations
of rows of A equal that row of E. One such solution can easily be found by
adding all rows of A less than or equal to the row of E in question. For low
dimensions some sort of branch-and-bound method can be used if all
possibildtiesare desired.
(3) The first row of G must be (0 0 1). The second row of G must be
(0 1 0). The th.’u d row of G must be (0 1 0). Thus G is
L 0
0 01 01
(4) For each chosen U, choose a vector s such that si = 1 only if A,* d fj
and si = 1 for at least one i such that A,. = o. We will sometimes write S(O)
for s.
(5) If it is not true that o’> u implies s(u’)= s(u). redefine s(u) to he
sO(z))+~z~O~O(z), where s”(u) is the s(z)) obtained in step (4).
(6) For a vector t which is not one of the chosen identification vectors,
if t is not less than any row basis vector, send t to an arbitrary sum of the
vectors s(u) or 0. Otherwise let p(t) be the infimum of the row vectors which
are greater than or equal to t, taken in the row space of A regarded as a
lattice. Choose a set P of row basis vectors whose sum is less than or equal to
p(t), and let t be sent to b=ZzErs(z), or if l? is empty, to 0.
(7) Write the vectors s and b in the order of the t’s and u’s.
EXAMPLE 3. Let A be
i 0111 1 01.
1 1
(1) b(A)= ((1 1 01,(0 1 1)).
(2) I((1 1 O))={(l 0 0)); I((0 1 l))={(O 0 1)).
(3) (Not needed.)
(4) For (1 1 0), s is (1 0 0). For (0 1 l), s is (0 0 1).
(5) (Not needed.)
(6) For t=(O 1 0), p(t)=O. So b-0.
(7) Thus the Vagner inverse of A is
6. ANALYSIS OF EFFICIENCY
greater than or equal to the t’s, requiring r(n - r) comparisons. For such a t
we next determine the set Ii of all row basis vectors less than or equal to
every row basis vector which is greater than or equal to t. This is done by
intersecting at most r columns in the matrix E. This requires at most
f(n - r)n steps for all the vectors t. Now r is chosen to be any subset of I,.
The sum required in computing z will be at most r(n - r)n separate addi-
tions. In general, to obtain a new Vagner inverse we need only perform one
of the operations under step (6), and many times step (5) need not be
performed. Step (7) requires at most n2 comparisons.
Thus our algorithm requires 2n3+4n2 - n initial steps and then k,n2 +
1.5n2+0.5n3 (or usually fewer) additional operations per Vagner inverse.
To find a single Vagner inverse we can choose si = 1 if and only if ri Q T,
so that step (5) will be automatic. Send all the t vectors to zero. In this way a
special Vagner inverse can be found with about 2n3 + 6n2 - n operations.
By comparison, Rao’s algorithm may require something like n3 steps to
find a row basis, n3 steps to find a column basis, and 2n2 steps of hunting for
columns with a single one entry, to obtain a g-inverse. Then somewhat more
steps may be needed (about n3) to obtain a Vagner inverse.
where S is the set of basis vectors which are not greater than or equal to v.
Thus there exists at least one identification vector for v if and only if
vg c w.
WES
256 K1 HANG KIM AND FRED W. ROUSH
Multiply both sic& .by v, in tne row space regarded as a lattice. That is
v=
(2 1WES
20 v.
By ciisrributimty,
Each vector WL)is properly less r&n 0, since each w was not greater than or
equal ro ~1.Thus v IS a sum of pop&y smaller vectors. Thus v is not a basis
vector, wnich is a conuadiction. Thrs proves the lemma. q
r~I’HEOM!M 2. lhe Jirst algtitiwn, fbr all allowable &&es, yi4dd.s all
g- imwrs~sof A.
Proof: First we will show that matrices G constructed in this way are
always g-mvtxses of A. ‘kt e7 be a row basis vector ior A. Then VG is the
sum of s(v), of Y(Q) for aU other basrs vectors v, whose identikicdtion vectors
iue b O, and or b(t) for all t< u. By the definition of s, sjv)A=u and
Y( SJA = v,. By the dehmtion of an identification vector, U, < v for each v.
And by deiimtion of b, b (t)A < y(t) and y(t) < v, for each t < v. Therefore
will ‘be a sum of vectors properly iess than v. Blur (v G)A = r). T’ms contradicts
the assumption that v is a basks vector. Thererore ror some mentnricauon
vector tl of v, there will be some i such that (uG), = I and Air- O. This
completes the proof of the thtxxem.
Proof. The mrmber of ways to map the set ot identiucauon vectors ior r)
subject to si = 1 only if Ai. < v and si = 1 for at Ieast one i sucfl mat A~, = u,
for some identification vector, is given by
25s KI HANG KIM AND FRED W. ROUSH
vGA=s(v)A+ 2 xs(z)A=v.
t<o r
Thus as before, for any vector va, vsAGA = v,A and AGA = A. The image of
G is generated by the vector s(v). We have s(v)A= v and
vG=+)f tFoFs(z,.
Each z in the second part of the sum is less than or equal to p(t) and so less
than or equal to v. Therefore by step (S), s(z) < s(v). Therefore vG=s(v). So
s(v)AG = s(v), and AG is the identity on the image of G. Therefore
GAG = G.
Next suppose G is a Vagner inverse of A. Since G will also be a g-inverse
of A, there must be an identification vector u mapped to vectors s(v) as
described in step (4) By the construction, s(v)A = v. Also VGA = v, since v is
in the image of A. Therefore s(v) = uG and vG are vectors in the image of G
which have the same image under A. But GAG = G implies A is l-l on the
image of G. Therefore s(v) = vG. Therefore v < v’, s(v) < s(v’).
Let L be the subspace spanned by the vectors s(v). We have s (v)A =
uGA = v, so L maps onto the space or row vectors of A, under A. Also L is
contained in the image of G. Since A is l-l on the image of G, L must be the
complete image of G, and A must be l-l on L.
Lettbeavectorwithexactlyonel,andletM={wER(A):w>t}.For
w E M, wGA= w, so tGA < w. Since tGA is in the row space of A,
INVERSES OF BOOLEAh’ MATRICES 259
tGA= &.
r
Since
If E is symmetric, so will be PAP”. Thus all the blocks below the main
diagonal wiU be zero as shown. This implies that any two row basis vectors
for PEPT have nonoverlapping sets of 1’s. The same must be true of E. If A
has the same row space as E, the sarle will be true of a row basis for A.
If A has this property, the matrix E such that ei i = 1 if and onIy if uJ = 1
and 4 = 1 for some basis vector o of A wiIl be a symmetric idempotent with
the same row space as A.
E is unique if E’ is another symmetric idempotent with the same row
space as A; E’ by symmetry must also have the same column space as A.
Therefore E, E’ lie in the same 3c-class [5]. But no X-class in any semigroup
can contain more than one idempotent [2]. This proves the theorem. I
Proof. This follows from the defining formulas AGA = A, (AG)?‘= A@.
Ip
[S(n,k)+(k+l)S(n,k+l)]~ (-1)‘(!)(2’-i)“.
i=l
Here the S (n, k) are the Stirling numbers of the seed kind.
$(-l)i(i)(2k-i)n
i=l
matrices with a particular choice of such row space. (This uses the fact that
the space generated by k vectors with disjoint sets of l’s has Zk elements.) n
INVERSES OF BOOLEAN MATRICES 261
PTOOf: *: Ry Theorem 7 the sets of I’s in any two row basis vectors are
disjoint, and by Theorem ‘7 applied to the transpose, the sets of l’s in any
two column basis vectors are disjoint. Suppose some row A,. is a nontrivial
sum invobing two row vectors A,,, A,, which are different basis vectors.
There will be some i, i such that (Aa*)i-l, (Ab.)i=O, (A,.)j=O, (Ab.)j=l.
Therefore there will be vectors v, w in the cdumn basis such that v, = 1,
u, = 0, w, = 0, w, = 1. IIowever a91column vectors w.hich are 1 in either the a
row or the b row will have a 1 in the c row. Thus v, w do not have disjoint
sets of l’s, a contradiction.
3: C. R. Rao [lo] has proved that if A has a Moore-Penrose inverse G, G
must be AT. We will show that AT is a Moore-Penrose inverse of A if A
satisfies the condition of this theorem.
The (i, /)-entry of XX TX is 1 if and only if there exist p, q such that
~~~=(X~)~~=x~~==l, or qp=xqp = xqi= 1. Therefore the i row and the q row
are not disjoint, so they must be equal, so xii = 1. Conversely if xii = 1, taking
p = i, y = i gives (XXTX)ii = 1. Therefore XX’X= X.
Transposing this equation, X ‘XX r = XT. The matrices XX and X TX will
always be symmetric. This proves the theorem. II
k![S(n,k)+(k+l)S(n,k+l)]“,
where the S (n, k) are the Stirling numbers of the second kind.
REFERENCES