Resolvable designs with large blocks
Resolvable designs with large blocks
Table 1
A resolvable design in D(9, 5; 5, 4)
1 6 1 2 1 4 1 3 1 2
2 7 4 3 2 5 2 4 3 5
3 8 5 6 3 6 5 6 4 6
4 9 7 8 7 8 8 7 7 9
5 9 9 9 8
2. Model, information and optimality criteria. Let yhjl denote the yield
from the lth experimental unit in block j of replicate h. Thus, the triples
(h, j, l) identify the experimental units and the design d corresponds to a map
d[h, j, l] from the units to the set of treatments. The standard linear model for
the yields incorporates a mean effect µ, replicate effects ρh , block effects βhj ,
treatment effects τd[h,j,l] and mean zero, uncorrelated, equivariable random
error terms ehjl :
yhjl = µ + ρh + βhj + τd[h,j,l] + ehjl ,
h = 1, . . . , r; j = 1, 2; l = 1, . . . , kj . This model may be written in matrix
terms as
y = µ1 + (Ir ⊗ 1v )ρ + Lβ + Xd τ + e,
4 J. P. MORGAN AND B. H. RECK
1
(4) ψE (zd ) = ,
zd1
or, equivalently,
n−l+1
X n−l+1
X
xi ≥ yi for all 1 ≤ l ≤ n,
i=n i=n
If the sequences {xi }ni=1 and {yi }ni=1 are written as the elements of vectors
x and y, then the fact that y majorizes x is written as y ≻ x or x ≺ y.
Let ed and ed∗ be the vectors of eigenvalues for the optimality matrices
for designs d and d∗ , respectively. Lemma 3 lists simple majorization facts
([3], page 30) used in subsequent sections. For majorization comparisons,
the third of these allows work directly with the edh rather than the r − kve1dh k2
(see Lemma 1). Thus, if ed ≻ ed∗ and the two vectors are not identical, then
d∗ is Schur-better than d. Design d∗ is Schur-optimal if ed ≻ ed∗ for every
d ∈ D.
Pn Pn
Lemma 3. For real numbers {xi }ni=1 and {yi }ni=1 with i=1 xi = i=1 yi ,
k2
Proof. Given the conditions on D, the inequalities k1 − k2 ≤ k1 +k 1
2
≤ k1
∗
imply that θ = θ is an admissible value for the common block concurrence
of an ECD(θ). For θ = θ ∗ , Md in (6) is vp I. Since the eigenvalues of Md are all
identical, they are majorized by the eigenvalues of every competing design.
Example. Consider the setting D(9, 4; 6, 3). Then (k1 + k2 )|k12 with θ ∗ =
4 and if an ECD(4) exists, it is Schur-optimal. In fact, an ECD(4) does exist
and is shown in Table 2.
The settings for which k12 is a multiple of k1 + k2 are relatively sparse
(a situation much like that of BIBDs relative to all simple block design
8 J. P. MORGAN AND B. H. RECK
Table 2
A Schur-optimal ECD( 4)
in D(9, 4; 6, 3)
1 7 1 5 1 3 3 1
2 8 2 6 2 4 4 2
3 9 3 9 5 9 5 9
4 4 6 6
5 7 7 7
6 8 8 8
settings). For the 1225 pairs 2 ≤ k2 < k1 ≤ 51, only 23 meet the divisibil-
ity requirement implied by (7). Theorem 4 is thus only a start, albeit an
important one.
3.2. Global optimality of other ECDs. Good designs are expected to have
eigenvalue structures “close” to that of ECD(θ ∗ )’s, suggesting this question:
Is some equal concurrence design Schur-optimal when (k1 + k2 )|/k12 ? To in-
vestigate this question, define the block concurrence parameter θ̄ by
k12
(8) θ̄ = int
k1 + k2
and write
k12
(9) γ= − θ̄.
k1 + k2
Then 0 ≤ γ < 1 and a necessary condition for existence of ECD(θ ∗ ) is γ = 0.
Consequently, γ is called the block discordancy coefficient; it measures the
departure of the block sizes from that required for equality of all eigenvalues.
The parameter γ will play a pivotal role in the remainder of this paper, as
will the concurrence discrepancies defined next.
Define the block concurrence discrepancy matrix ∆d = (δdhh′ ), where
if h 6= h′ ,
φdhh′ − θ̄,
δdhh′ =
0, if h = h′ .
For h 6= h′ , the off-diagonal elements δdhh′ will be referred to as block con-
currence discrepancies. Rewritten in terms of block discrepancies and the
discordancy coefficient, the general optimality matrix (5) becomes
p
(10) Md = I − γ(J − I) + ∆d .
v
For ECD(θ ∗ ), γ = 0, ∆d = 0 and Md = vp I. If γ 6= 0, (10) shows that the form
of optimal design may depend both on the magnitude of γ and the values
RESOLVABLE DESIGNS 9
of the discrepancies. Matrix (10) is completely symmetric for, and only for,
ECDs, in which case ∆d is an integer multiple of J − I.
The ECDs which are combinatorially closest to ECD(θ ∗ ) are those with
either θ = θ̄ or θ = θ̄ + 1. Not surprisingly, these are strong competitors in the
optimality race. ECD(θ̄)s have φdhh′ = θ̄ for each 1 ≤ h 6= h′ ≤ r, so ∆d = 0
and the eigenvalues of Md are
p p
(11) ξ1 (γ) = + γ and ξ2 (γ) = − (r − 1)γ,
v v
with frequencies r − 1 and 1, respectively. The two eigenvalues satisfy ξ1 (γ) ≥
ξ2 (γ). If φdhh′ = θ̄ + 1 for every h 6= h′ , then ECD(θ̄ + 1)’s have ∆d = (J − I)
and the eigenvalues of Md are
p
ξ1 (γ − 1) = − (1 − γ) and
v
(12)
p
ξ2 (γ − 1) = + (r − 1)(1 − γ),
v
with frequencies r − 1 and 1, respectively, and with ξ2 (γ − 1) ≥ ξ1 (γ − 1).
2vr v2
= (v − 1)r 2 − tr Md + 2 tr Md2
p p
2v 2 X X
= (v − 3)r 2 + r + (δdhh′ − γ)2 ,
p2 h<h′
A result of a similar flavor holds for ECD(θ̄ + 1)’s using the E-criterion.
As pointed out by Kunert [15], page 385, designs with eigenvalues zdi in the
form of Lemma 3(ii) are Schur-best whenever they are E-optimal. For the
current problem, this is stated as follows.
Though design nonexistence can play a role, Corollary 8 says that one
does not expect to find a Schur-optimal design for 1r ≤ γ ≤ r−1
r (Theorem 5
makes for an interesting juxtaposition). Schur-domination can nonetheless
be used to eliminate many competitors, as shown next.
RESOLVABLE DESIGNS 11
Proof. For the UCD d as described in the corollary, take δd12 ≤ −α.
Then from Lemma 9, ed1 ≥ pv + γ − α and vp − γ + α ≥ edr . By Theorem 6,
an ECD(θ̄) is Schur-better than d if ξ2 (γ) ≥ pv − γ + α ≥ edr ⇐⇒ γ ≤ r−2
α
.
p
By Theorem 7, an ECD(θ̄ + 1) is Schur-better than d if ed1 ≥ v + γ − α ≥
ξ2 (γ − 1) ⇐⇒ γ ≥ r−α−1
r .
Corollary 11. When r ≤ 4, all UCDs with some δdhh′ ≤ −1 are Schur-
inferior to an ECD, and when r = 5 or 6, UCDs with some δdhh′ ≤ −2 are
Schur-inferior to an ECD.
Corollary 13. When r ≤ 4, all UCDs with some δdhh′ ≥ 2 are Schur-
inferior to an ECD, and when r = 5 or 6, UCDs with some δdhh′ ≥ 3 are
Schur-inferior to an ECD.
Corollaries 11 and 13 say that optimal (with respect to any convex cri-
terion) designs in settings D(v, r; k1 , k2 ) with r ≤ 4 must be an ECD(θ̄), an
ECD(θ̄ + 1) or an AECD. Optimal designs in settings with r = 5 or 6 must
have block concurrence discrepancies δdhh′ ∈ {−1, 0, 1, 2} for all 1 ≤ h 6= h′ ≤
r. It is unlikely that such global statements can be much improved. The im-
portance of these results lies in the prevalence of small r in the application
of resolvable designs.
Lemma 17. For d ∈ D(v, r; k1 , k2 ) and with P = (I − 1r J), let ud1 and udr
be the maximum and minimum eigenvalues of P ∆d P , respectively. Then:
14 J. P. MORGAN AND B. H. RECK
p
(i) if ud1 > 0 then ed1 ≥ v + γ + ud1 ;
(ii) edr ≤ vp + γ + udr .
Parts (i) and (ii) of Corollary 18 provide alternative routes (cf. Corollar-
ies 10 and 12) for establishing Schur-domination of ECDs. Part (iii) is the
key part of the proof of Theorem 14. If d is not to be eliminated by ECD(θ̄),
then P ∆d P cannot have a positive eigenvalue, that is, must be nonpositive
definite. Recall that the current E-competitors are all d for which ∆d is the
adjacency matrix of a simple, undirected graph.
Lemma 19. Let A be the adjacency matrix for a simple undirected graph
of r vertices. PAP is nonpositive definite if and only if A − J may be writ-
ten (possibly after vertex permutation) in the form (13) for some positive
integers n and t1 ≤ t2 ≤ · · · ≤ tn with ni=1 ti = r.
P
1/2 1/2
Consequently, Hd is nonnegative definite if and only if Dt EDt is non-
negative definite. This is so if and only if E is nonnegative definite, that is,
if and only if 1 − γ ≤ n1 . Thus, for any γ ≤ r−1 r , the E-optimal designs are
all group-affine designs for which the number of groups n is no larger than
1
1−γ , that is, for which n ≤ nγ . This completes the proof of Theorem 14.
To prove Theorem 15, recall the definition of t(d) given just prior to Theo-
rem 14 and note that now only d ∈ DE is being considered. It is shown in Ap-
pendix A.4 that the eigenvalues of Hd , aside from r − nγ zeros, are the eigen-
values of E 1/2 Dt E 1/2 . Thus, the problem is to show that t(d1 ) ≺ t(d2 ) implies
the eigenvalues of E 1/2 Dt(d1 ) E 1/2 are majorized by those of E 1/2 Dt(d2 ) E 1/2 .
Now, t(d1 ) ≺ t(d2 ) is Pequivalent to t(d1 ) = St(d2 ) , where S is a doubly
stochastic matrix; S = m i=1 ai Qi for permutation matrices Qi and positive
numbers ai summing to 1 ([3], pages 33 and 37). Thus,
E 1/2 Dt(d1 ) E 1/2 = E 1/2 DSt(d2 ) E 1/2
" #
= E 1/2 ai Qi Dt(d2 ) Qi E 1/2
X
and Kageyama [25] and Ghosh and Teschmacher [10]. Here, only strength-
two arrays on two symbols will be needed. A balanced array of strength 2,
BA(N, m, 2), on the symbols 0 and 1, is an N × m array with the property
that for any selection of two columns, the N pairs formed by the rows are
(0, 0), (0, 1), (1, 0) and (1, 1), with frequencies µ0 , µ1 , µ1 and µ2 , respec-
tively. Two-symbol orthogonal arrays of strength two are the special case
µ0 = µ1 = µ2 . When the number of 0’s in each column is specified (as is
the case below), the µi ’s are all determined by θ = µ0 and the array will be
denoted BA(N, m, 2; θ).
Bailey, Monod and Morgan [1] demonstrate the combinatorial equivalence
between affine resolvable designs and orthogonal arrays. Their method can
be used to express any resolvable design as a combinatorial array, as follows.
Given a resolvable design for v treatments in r replicates, each consisting of
s blocks, label the blocks within a replicate 0, . . . , s − 1. Construct a v × r
array by identifying rows of the array with treatments of the resolvable
design, and columns with replicates: symbol j ∈ {0, 1, . . . , s − 1} is placed
in row i, column h if and only if treatment i is in block j of replicate h.
Theorem 24 is evident.
row has 1 in the first v+22 columns. Delete the first two rows and the
first v+2
2 columns, then replace −1 by 0 throughout. Clearly, the result is
v v−2
BA(v, 2 , 2; 4 ).
APPENDIX A: PROOFS
1/2
A.1. Proof of Lemma 1. Let Ds be the diagonal matrix of square roots
−1/2
of block sizes and write Bd = Nd Ds . Multiplying Cd by 1r and right and
−1/2
left multiplying Cdual by Ds , equations (1) and (2) become
1 1 1
(16) Cd = I − Nd Ds−1 Nd′ = I − Bd Bd′ = Cd∗
r r r
and
1 1
(17) Ds−1/2 Cdual Ds−1/2 = I − Ds−1/2 Nd′ Nd Ds−1/2 = I − Bd′ Bd = Cdual
∗
.
r r
Since (1) and (16) differ only by a constant and since the nonzero eigenvalues
of Bd Bd′ and Bd′ Bd are identical, the eigenvalues of Cd can be found from
∗
those of Cdual in (17), which depends on block concurrences and block sizes.
−1/2 −1/2
∗
Continuing, write Cdual = I − 1r Ad for Ad = Bd′ Bd = Ds Nd′ Nd Ds .
1/2 1/2
Regardless of the design d, Ad Ds 1 = rDs 1; r is an eigenvalue of Ad
corresponding to the zero eigenvalue common to Cdual and Cd . One term of
the spectral decomposition of Ad is then
1/2 1/2 √
r(Ds 1)(Ds 1)′ 1
(18) = J⊗ √ k1 k1 k2
,
1/2 ′ 1/2
(Ds 1) (Ds 1) (k1 + k2 ) k1 k2 k2
A.2. Proof of Lemma 17. The lower bound on ed1 follows from
ed1 = max
′
x′ Md x
x x=1
p
′
= max x + γ I − γJ + ∆d x
′
x x=1 v
p
′
≥ max x + γ I − γJ + ∆d x
x′ x=1 v
x′ 1=0
p
= + γ + max x′ ∆ d x
v x′ x=1
x′ 1=0
p
= + γ + max x′ P ∆ d P x
v x′ x=1
x′ 1=0
p
= + γ + max x′ P ∆ d P x
v x′ x=1
p
= + γ + ud1 .
v
The penultimate equality holds since ud1 > 0 and 1 is an eigenvector of
P ∆d P with eigenvalue 0. Likewise,
edr = min
′
x′ Md x
x x=1
p
′
= min x + γ I − γJ + ∆d x
′ x x=1 v
p
≤ + γ + min x′ P ′ ∆ d P x
v x′ x=1
x′ 1=0
p
= + γ + udr .
v
The last equality holds provided udr < 0, for reasons similar to those above.
If udr > 0, the bound still holds, since edr ≤ tr(M
r
d)
= vp ≤ vp + γ + udr .
= max y ′ (A − J)y
y=P x,x′ x=1
t t
(i)
!2
X X
= max − yj ,
y=P x,x′ x=1
i=1 j=t(i−1) +1
22 J. P. MORGAN AND B. H. RECK
M3 = (−1) 0 1 1 ,
1 1 1
which is not of the form (13), contradicting the fact that the result holds
for r = 3. Thus, at most one of a1 , a2 is nonzero. It follows immediately
that for n > 2, the same statement holds for any pair ai , ai′ . Permuting
rows and columns as needed, it can be assumed that ai = 0ti for i < n and
consequently that Ar+1 − Jr+1 has the form (13).
A.4. Eigenvalue equations for Hd and proof of Lemma 20. Any vector of
the form (c′1 , c′2 , . . . , c′n ), where ci ∈ ℜti is either a contrast vector or the zero
vector, is an eigenvector of Hd with eigenvalue 0. Consequently, all other
eigenvectors are of the form e′ = (x1 1′t1 , x2 1′t2 , . . . , xn 1′tn ) for some scalars
x1 , . . . , xn . The first equation in the system Hd e = λe is γt1 x1 − (1 − γ)t2 x2 −
· · · − (1 − γ)tn xn = λx1 ; the other equations may be written similarly in
order to see that Hd e = λe are equivalent to [Dt − (1 − γ)1n t′ ]x = λx, where
RESOLVABLE DESIGNS 23
the last step following because GG′ and G′ G have the same eigenvalues for
any square G. Now, in the proof of Theorem 15 some elements of t are
allowed to be zero (without loss of generality, t1 = · · · = tz = 0 for integer
z ≥ 1), in which case Dt is not invertible. In this case there are z additional
zero eigenvalues plus a reduced system of n − z equations in tz+1 , . . . , tn . It
1/2 1/2 1/2 1/2
is easy to see that |Dt EDt − λI| = |0z,z − λIz ||D̃t Ẽ D̃t |, where Ẽ and
D̃t are the lower-right submatrices of E and Dt of order n − z. Thus, the n
1/2 1/2
eigenvalues sought are still those of Dt EDt and thus of E 1/2 Dt E 1/2 .
11111 11111
00000 11111
01001 01100
01010 00010
00110 10010
10010 00001
00101 10001
11100 00110
10001 10100
00101
01010
01001
11000
24 J. P. MORGAN AND B. H. RECK
REFERENCES
[1] Bailey, R. A., Monod, H. and Morgan, J. P. (1995). Construction and optimality
of affine-resolvable designs. Biometrika 82 187–200. MR1332849
[2] Bailey, R. A. and Speed, T. P. (1986). Rectangular lattice designs: Efficiency
factors and analysis. Ann. Statist. 14 874–895. MR0856795
[3] Bhatia, R. (1997). Matrix Analysis. Springer, New York. MR1477662
[4] Bose, R. C. (1942). A note on the resolvability of balanced incomplete block designs.
Sankhyā 6 105–110. MR0008064
[5] Chakravarti, I. M. (1956). Fractional replication in asymmetrical factorial designs
and partially balanced arrays. Sankhyā 17 143–164. MR0085677
[6] Chakravarti, I. M. (1961). On some methods of construction of partially balanced
arrays. Ann. Math. Statist. 32 1181–1185. MR0130770
RESOLVABLE DESIGNS 25