Fundamental Concepts in The Design of Experiments 3nbsped 0030617065 Compress
Fundamental Concepts in The Design of Experiments 3nbsped 0030617065 Compress
V43
l^iMiM
bsKt'P^fi ipipl>J:.-|
M
Jy;j i
^1 i|;i sMi
Digitized by the Internet Archive
in 2017 with funding from
China-America Digital Academic Library (CADAL)
https://archive.org/details/fundamentalconceOOhick_0
/
/•
■*
.4
■-#
vW
• a
4
-i*'"
•'*. ■« ,
H.
r
Fundamental
Concepts in
the Design of
Experiments
Fundamental
Concepts in
the Design of
Experiments
Third Edition
CHARLES R. HICKS
Professor of Statistics and Education
Purdue University
Bibliography: p. 377
Includes index.
1. Experimental design. I. Title.
QA279.H53 1982 001.4'34 82-12062
ISBN □-□3-Dtil70b-S
/
The Experiment, the Design, and the Analysis 1
1.1 INTRODUCTION TO EXPERIMENTAL DESIGN 1
1.2 THE EXPERIMENT 2
1.3 THE DESIGN 4
1.4 THE ANALYSIS 6
1.5 SUMMARY IN OUTLINE 6
1.6 EXAMPLES 7
PROBLEMS 14
2
Review of Statistical Inference 15
2.1 INTRODUCTION 15
2.2 ESTIMATION 17
2.3 TESTS OF HYPOTHESES 19
2.4 THE OPERATING CHARACTERISTIC CURVE 21
2.5 HOW LARGE A SAMPLE? 25
2.6 APPLICATION TO TESTS ON VARIANCES 25
2.7 APPLICATION TO TESTS ON MEANS 27
2.8 APPLICATION TO TESTS ON PROPORTIONS 30
PROBLEMS 32
V
vi Contents
4
Single-Factor Experiments—Randomized Block and
Latin Square Designs 66
4.1 INTRODUCTION 66
4.2 RANDOMIZED COMPLETE BLOCK DESIGN 67
4.3 ANOVA RATIONALE 72
4.4 MISSING VALUES 73
4.5 LATIN SQUARES 76
4.6 INTERPRETATIONS 78
4.7 GRAECO-LATIN SQUARES 79
4.8 EXTENSIONS 80
4.9 SUMMARY 80
PROBLEMS 80
5
Factorial Experiments 86
5.1 INTRODUCTION 86
5.2 FACTORIAL EXPERIMENTS 91
5.3 INTERPRETATIONS 94
5.4 COMPUTER PROGRAM 96
5.5 ANOVA RATIONALE 98
5.6 REMARKS 103
5.7 SUMMARY 104
PROBLEMS 105
6
2^ Factorial Experiments 112
6.1 INTRODUCTIONf 112
6.2 2^ FACTORIAL 112
6.3 2^FACTORIAL 120
6.4 2^—REMARKS 123
6.5 SUMMARY 124
PROBLEMS 125
7
Qualitative and Quantitative Factors 129
7.1 INTRODUCTION 129
7.2 LINEAR REGRESSION 130
7.3 CURVILINEAR REGRESSION 137
7.4 ORTHOGONAL POLYNOMIALS 139
7.5 COMPUTER PROGRAM 144
7.6 QUALITATIVE AND QUANTITATIVE FACTORS 144
7.7 TWO FACTORS—ONE QUALITATIVE, ONE QUANTITATIVE 145
7.8 TWO FACTORS—BOTH QUANTITATIVE 152
Contents
a
Multiple Regression 164
8.1 INTRODUCTION 164
8.2 THE MULTIPLE REGRESSION MODEL 165
8.3 COMPUTER PROGRAMS 175
8.4 REMARKS 176
PROBLEMS 177
9
3^ Factorial Experiments 187
9.1 INTRODUCTION 187
9.2 3^ FACTORIAL 187
9.3 3^ FACTORIAL 198
9.4 SUMMARY 207
PROBLEMS 207
lO
Fixed, Random, and Mixed Models 210
10.1 INTRODUCTION 210
10.2 SINGLE-FACTOR MODELS 211
10.3 TWO-FACTOR MODELS 212
10.4 EMS RULES 214
10.5 EMS DERIVATIONS 218
10.6 THE PSEUDO-F TEST 221
10.7 REMARKS 223
PROBLEMS 224
II
Nested and Nested-Factorial Experiments 227
11.1 INTRODUCTION 227
11.2 NESTED EXPERIMENTS 228
11.3 ANOVA RATIONALE 232
11.4 NESTED-FACTORIAL EXPERIMENTS 233
11.5 COMPUTER PROGRAMS 238
11.6 REPEATED-MEASURES DESIGN AND NESTED-FACTORIAL
EXPERIMENTS 239
11.7 SUMMARY 244
PROBLEMS 244
14
Factorial Experiment—Confounding in Blocks 281
14.1 INTRODUCTION 281
14.2 CONFOUNDING SYSTEMS 282
14.3 BLOCK CONFOUNDING—NO REPLICATION 285
14.4 BLOCK CONFOUNDING WITH REPLICATION 290
14.5 SUMMARY 299
PROBLEMS 300
15
Fractional Replication 303
15.1 INTRODUCTION 303
15.2 ALIASES 304
15.3 FRACTIONAL REPLICATIONS 306
15.4 SUMMARY 314
PROBLEMS 315
lO
Miscellaneous Topics 318
16.1 INTRODUCTION 318
16.2 COVARIANCE ANALYSIS 318
16.3 RESPONSE-SURFACE EXPERIMENTATION 328
16.4 EVOLUTIONARY OPERATION (EVOP) 341
16.5 ANALYSIS OF ATTRIBUTE DATA 351
16.6 RANDOMIZED INCOMPLETE BLOCKS—RESTRICTION ON
EXPERIMENTATION 355
16.7 YOUDEN SQUARES 360
PROBLEMS 362
17
Summary and Special Problems 366
' » 4
t y '»' V' ■ ‘ V
4
'■^* ..f,. -. .■/ «>•
•s V 1
. - / JV, /' * •
.
■ "'■ ■ .V-, - ■'•- ' V'-';T>;- • ■ ' ' " '» ’ ^
^ V
W‘
. :■ 4 ;' V
■* i tw--
•'!,
i
H>
Vf
r^- ■-'•■
iTj?.. ^ ■> --. y
-f ?»)
. • V'
. '1 .
•* •
V
i-rr "}*
4-* *- . ; '
1 ^
fS ■ -St
. V -■ ■ I’ ■
«■' :'•» T
.i •
Kt "I
1 *' t’*' 1 ■■ i.*i’ V '' ■■ -:i-*
-I *
, V'..ii ’ ** •i - ■'
iV' I , I ‘ ‘i >5i ^ .f /■-!» ■ ■'■< :.<*
m.' '>.1 ’■f ■: -
Hp< a
■'■ K . ?j#.'
“•It ■ "■
f■ - i .:«•
I
:^V ^ ^ ••■*'
I * .■ »■? ' - V': ‘
f • < f
■-v»
■t.V
f.. s*'- • .
.■
^ * 1 i
^1.
'{♦ ■ ».
»• t ^s',1 t -^^.V k
««
. ' f 1^^'* TUN'* ' ** , -::
V.... i - 1 ' .' o. •
■JU. ►
■ ■ . 4.*'> -W > • ' >t< > V• » ' ,
r-v ^. <■ ‘ .
- - •'■ i ^ '• , ~ ,j .Z S \^*V'■ ,'
■ • .'v- V'
- t ‘ ■ '
^■■V . J ?•' .'jy ... -'V .■ ^ '
/ { gt ;fi 11
'y/'*A •-A ’ * / ^ '* 1 '' '
K
■ ■> f i ; i "\i
T: ■ ■’ ■>•
. i. ■■ i *i
'.
I, -* ■ '.Y(
• 1' >
n,
•A-
I 4
t *
» .. I
'V .
A • ' . ■■*■ l» : •) '
V,. 1* ■■
I'. '•; ^ ■, ^ na ' t A-'*'
“a
P.'
»' ) t:r7
i Q. *
|V
. •
' 1
I “8|
One of the main reasons for a third edition has been the use of computers to
simplify the analysis of data in an experiment. Computer programs follow most of
the complex examples showing the input instructions to a computer, the printout
from the computer, and the interpretation of the results.
The introduction has been completely revised to set the stage for ex-
perimental research while still maintaining the general theme of experiment,
design, and analysis. Chapters 4 and 5 of the previous edition have been combined
as they both represent restrictions on randomization, and a new chapter on
multiple regression has been added. Some topics rarely taught in a single-semester
course have been relegated to the chapter on miscellaneous topics.
Many illustrative examples have been replaced with examples that are more
practical and better illustrate the technique. There are 60 percent more problems
than in the second edition, resulting in more than 300 practical problems. Chapter
17 presents ten somewhat challenging problems for class discussion.
Two minor points: (1) Factorials have been introduced as 2^ and 3^ instead of
as 2" and 3", where/represents the number of factors in the experiment and n is
reserved for the number of observations per treatment. (2) The use of single and
double horizontal and vertical lines to show how the randomization has been
restricted should help illustrate how data were collected.
I am indebted to McElrath and Associates, Inc. (MAI), Management Consul-
tants, Minneapolis, Minnesota, for the opportunity to be exposed to many practic-
al problems in various industries over the past 20 years. Many of the text examples
have come from MAI clients. Several examples, especially in the field of educa-
tion, have come from students’ master’s and doctoral research. I want to thank the
following students for examples taken from their theses: Roger Bollenbacher,
Robert Horn, James Lehman, Willard Lewallen, Charlotte Macy, Thomas Mar-
shall, Larry Miller, Barbara O’Conner, Sallam Sallam, Nadia Sherief, Ruth
Tanner, John Wetz, and Catherine Yeotis.
xi
xii Preface to the third edition
Last, and most significant, I wish to thank Purdue University for granting me
sabbatical leave to complete this third revised edition.
Charles R. Hicks
Lafayette, Indiana
January, 1982
Preface to the Second Edition
At first glance the reader may not notice much change in this edition from the
original. An attempt has been made to keep chapter and section headings the same.
It cannot be considered as a major revision although there are many new features.
The new material adds approximately 25% to the size of the book and the number
of problems has been incrased by 70%.
Main features include a section on covariance analysis and analysis of
attribute data in the 16th chapter, introduction of linear and curvilinear regression
in Chapter 8, and an amplification of the concept of orthogonal contrasts and
orthogonal polynomials. Some topics have been rearranged to appear earlier in the
text and methods of testing means after analysis of variance have been changed
and extended.
Several users of the original text have sent me helpful suggestions for this
revision. Whenever a suggestion came from more than one person, I attempted to
include it in this edition. Naturally, all suggestions were not followed and the
responsibility for their adoption is mine alone. I am indebted to the following
people who made detailed suggestions:
V. L. ANDERSON Purdue University
O. A. BECKENDORF General Motors Institute
H. C. CHARBOMEAU General Motors Institute
R. I. FIELDS University of Louisville
R. J. FOPMAN University of Cincinnati
B. HARSHBARGER Virginia Polytechnic Inst.
O. B. MOAN Arizona State University
R. G. MORRIS General Motors Institute
V. V. NAMILOV MOSCOW State University
J. B. NEUHARDT The Ohio State University
C. R. PETERSON General Motors Institute
B. T. RHODES, JR. University of Houston
H. W. STOKER The Florida State University
xiv Preface to the second edition
Charles R. Hicks
Lafayette, Indiana
March, 1973
Preface to the First Edition
It is the primary purpose of this book to present the fundamental concepts in the
design of experiments using simple numerical problems, many from actual re-
search work. These problems are selected to emphasize the basic philosophy of
design. Another purpose is to present a logical sequence of designs that fit into a
consistent outline; for every type of experiment, the distinctions among the
experiment, the design, and the analysis are emphasized. Since this theme of
experiment-design-analysis is to be highlighted throughout the text, the first
chapter presents and explains this theme. The second chapter reviews statistical
inference, and the remaining chapters follow a general outline, with an experi-
ment-design-analysis summary at the end of each chapter.
The book is written for anyone engaged in experimental work who has a good
background in statistical inference. It will be most profitable reading to those with
a background in statistical methods including analysis of variance.
Some of the latest techniques have not been included. The reason for this is
again the purpose of presenting the fundamental concepts in design to provide a
background for reading journals in the field and understanding some of the more
recent techniques. Although it is not assumed that every reader of this book is
familiar with the calculus, some theory is presented where calculus is used. The
reader should be able to grasp the underlying ideas of each chapter even if he must
omit these sections.
This book evolved from a course in Design of Experiments given at Purdue
University over the past four or five years. The author is indebted to Professors
Gayle McElrath of the University of Minnesota and Clyde Kramer of Virginia
Polytechnic Institute whose lectures, along with the author’s, form the basis of this
book. Credit is also due Mr. James Dick and Mr. Roland Rockwell, who took
notes of the author’s lectures.
I shall be forever indebted to the University of Florida for granting me a
visiting professorship during the academic year 1962-1963 with time to work on
XV
xvi Preface to the first edition
this book, and to Mrs. Alvarez of their Industrial Engineering Department for
typing the manuscript.
Charles R. Hicks
Lafayette, Indiana
November, 1963
Fundamental
Concepts in
the Design of
Experiments
1
The Experiment,
the Design,
and the Analysis
1
2 The experiment, the design, and the analysis
Factor A
1 2 3
1
Factor B
2
Supplier
1 2 3
This involves six sets of data as did the above factorial, but now the factors
are in a decidedly different arrangement.
All of these points should be considered as part of the experiment phase
of any research.
The final step, analysis, includes the procedure for data collection, data
reduction, and the computation of certain test statistics to be used in making
decisions about various aspects of an experiment. Analysis involves the
computation of test statistics such as t, F, and their corresponding decision
rules for testing hypotheses about the mathematical model. Once the test
statistics have been computed, decisions must be made. These decisions
should be made in terms that are meaningful to the experimenter. They
should not be couched in statistical jargon such as “the third-order A x
B X E interaction is significant at the 1 percent level,” but instead should
be expressed in graphical or tabular form, in order that they be clearly
understood by the experimenter and by those persons who are to be “sold”
by the experiment. The actual statistical tests should probably be included
only in the appendix of a report on the experiment. These results should also
be used as “feedback” to design a better experiment, once certain hypotheses
seem tenable.
I. Experiment
A. Statement of problem
B. Choice of response or dependent variable
C. Selection of factors to be varied
D. Choice of levels of these factors
1. Quantitative or qualitative
2. Fixed or random
E. How factor levels are to be combined
II. Design
A. Number of observations to be taken
B. Order of experimentation
C. Method of randomization to be used
D. Mathematical model to describe the experiment
E. Hypotheses to be tested
III. Analysis
A. Data collection and processing
B. Computation of test statistics ,
C. Interpretation of results for the experimenter
1.6 Examples 7
1.6 EXAMPLES
Yij - Tj
HQ : P\ = p'2
with alternative
^ 1 • P'l < P'l
where the sign < indicates that we are interested in showing that the true
proportion contracting polio will be less for treatment 1 (vaccine) than for
1.6 Examples 9
treatment 2 (placebo). The primes on the p values are to show that the
hypotheses are statements about population proportions whereas the sam-
ples will give Pi and P —observed sample proportions. The data are shown
2
in Table 1.1.
Number
Sample Contracting Proportion
Treatment Size Polio Contracting
Pi - Pi
^pq(l/ni + I/MJ)
children. Here q = i — p.
If one now assumes an a risk of 0.001—taking one chance in 1000 of
rejecting HQ when it is true—the rejection region for z is given by z < — 3.09
(from the normal distribution table). Substituting the values above:
28 X 10“^ - 71 X 10"^
749 X 10^^(1 - 49 X 10“^)[l/200,745 + 1/201,229]
z = —6.14
which gives a very, very strong indication that we should reject HQ and
conclude that the vaccine was indeed effective.
In fact, the probability of getting a z as low as or lower than —6.14
when there is really no difference in the two groups is less than one chance
in a billion. These amazing results, of course, have been strongly substan-
tiated by the almost complete eradication of polio in the United States.
Example 1.2 The following example [18] is presented to show the three
phases of the design of an experiment. It is not assumed that the reader is
familiar with the design principles or analysis techniques in this problem.
The remainder of the book is devoted to a discussion of many such principles
and techniques, including those used in this problem.
An experiment was to be designed to study the effect of several factors
on the power requirements for cutting metal with ceramic tools. The metal
10 The experiment, the design, and the analysis
Tool Type T
1 2
Bevel B Bevel B
Type Cut
of C
2 2
3
4
Tool Type T
1 :1
Totals 31 108.93
* One asterisk indicates significance at the 5 percent level; two, the 1 percent level; and three,
the 0.1 percent level. This notation will be used throughout this book.
square of 2.23 with 24 degrees of freedom (df). The proper test statistic is
the F statistic (Appendix, Table D) with 1 and 24 df. At the 5 percent signifi-
cance level (a = 0.05) the critical region of F is F > 4.26. Comparing each
mean square with the error mean square indicates that only two hypotheses
can be rejected: bevel has no effect on deflection and type of cut has no
effect on deffection. None of the other hypotheses can be rejected, and it is
concluded that only the angle of bevel and type of cut affect power consump-
tion as measured by the y deffection on the dynamometer. Tool type appears
to have little effect on the y deflection, and all interactions are negligible.
1.6 Examples 13
Calculations on the original data of Table 1.2 show the average y deflections
given in Table 1.4.
These averages seem to bear out the conclusions that bevel affects y
deflection, with a 30° bevel requiring more power than the 15° bevel (note
the difference in average y deflection of 1.6 mm), and that type of cut affects
y deflection, with a continuous cut averaging 2.0 mm more deflection than
an interrupted cut. The difference of 0.6 mm in average deflection due to
tool type is not significant at the 5 percent level. Graphing all four B-C
combinations in Figure 1.2 shows the meaning of no significant interaction.
Y
y Deflection
(mm)
easily be seen that an increase in the degree of bevel produced about the
same average increase in y deflection regardless of which type of cut was
made. This is another way to interpret the presence of no interaction.
The experiment here was a three-factor experiment with two levels for
each factor. The design was a completely randomized design and the analysis
was a three-way ANOVA with four observations per cell. From the results
of this experiment the experimenter not only found that two factors (angle
of bevel and type of cut) affect the power requirements, but also determined
that within the range of the experiment it makes little difference which
ceramic tool type is used and that there are no significant interactions
among the three factors.
These are but two examples that might have been used to illustrate the
three phases of the project: experiment, design, and analysis.
PROBLEMS
1.1 In your own area of interest, write out the statement of a problem that you believe
is researchable.
1.2 For the problem stated in (1), list the dependent variable or variables of interest
and how you propose to measure them.
1.3 For the problem stated in (1), list the independent variables that might affect your
dependent variable(s) and explain how you would handle each variable.
1.4 Prepare a data layout for collecting data on your problem.
2
Review of
Statistical Inference
2.1 INTRODUCTION
In any experiment the experimenter is attempting to draw certain inferences
or make a decision about some hypothesis or “hunch” concerning the
situation being studied. Life consists of a series of decision-making situations.
As you taste your morning cup of coffee, almost unconsciously you decide
that it is better than, the same as, or worse than the coffee you have been
drinking in the past. If, based on your idea of a “standard” cup of coffee, the
new cup is sufficiently “bad,” you may pour it out and make a new batch.
If you believe it superior to your “standard,” you may try to determine
whether the brand is different than the one you are used to or the brewing
is different, and so forth. In any event, it is the extreme differences that may
cause you to decide to take action. Otherwise, you behave as if nothing has
changed. In any decision you run the risk that the decision is wrong since
it is based on a small amount of data.
In deciding whether or not to carry an umbrella on a given morning, you
“collect” certain data: you tap the barometer, look at the sky, read the
newspaper forecast, listen to the radio, and so on. After quickly assimilating
all available data—including such predictions as “a 30 percent probability
of rain today”—you make a decision. Somehow a compromise is made
between the inconvenience of carrying the umbrella and the possibility of
having to spend money for a cleaning bill for one’s clothes.
In these instances, and in most everyday events, decisions are made in
the light of uncertainty. Statistics may be defined as a tool for decision making
in the light of uncertainty. Uncertainty does not imply no knowledge, but
only that the exact outcome is not completely predictable. If ten coins are
tossed, one knows that the number of heads will be some integer between
0 and 10. However, each specific integer has a certain chance of occurring
and various results may be predicted in terms of their chance or probability
of occurrence. When the results of an experiment that are observed could
have occurred only 5 times in 100 by chance alone, most experimenters
15
16 Review of statistical inference
consider that this is a rare event and will state that the results are statistically
significant at the 5 percent significance level. In such cases the hypothesis
being tested is usually rejected as untenable. When statistical methods are
used in experimentation, one can assess the magnitude of the risks taken in
making a particular decision.
Statistical inference refers to the process of inferring something about a
population from a sample drawn from that population. The population
consists of all possible values of some random variable Y, where Y represents
the response or dependent variable being measured. The response Y may
represent tensile strength, weight, score, reaction time, or whatever criterion
is being used to evaluate the experimental results. Characteristics of the
population of this random variable are called parameters. In general 6 will
be used to designate any given population parameter. The average or ex-
pected value of the random variable is designated as E{Y) = p. If the
probability function defining the random variable is known,
E(Y) = Y. Y^PiYi)
where p(Yi) is a discrete probability function, and
E{Y) = jYf(Y)dY
where n is the number of observations in the sample, and the sample variance
= t (Yi - Yf/in - 1)
i= 1
Italic letters will be used to designate sample statistics. The symbol u will
be used to designate a general statistic corresponding to the population
parameter 6.
Most statistical theory is based on the assumption that samples drawn
are random samples, that is, that each member of the population has an
equal chance of being included in the sample and that the pattern of variation
in the population is not changed by this deletion of the n members for the
sample.
2.2 Estimation 17
The notion of statistical inference may be divided into two parts: (1)
estimation and (2) tests of hypotheses.
2.2 ESTIMATION
Note that the sum of squares Yj= i (F ~ F)^ must be divided by n and
not by n if s^ is to be unbiased. The sample standard deviation
s Yf/in - 1)
This somewhat subtle point is proved in Burr (see [7], pp. 101-104).
A few basic theorems involving expected values that will be useful later
in this book include
E{k) = k (where /c is a constant)
E{kY) = kE{Y)
E{Y, + Y2} = E(YY E{Y2)
E{Y - PYY = E{Y^) - pj = al
E — {n — 1)<7Y
or
£(SSy) = {n - 1)(7?
or
£(SS J = (df on SS„)(T„^
Y ± 1.96
where 1.96 is taken from normal distribution values (see Appendix, Table A),
n is the sample size, and a is the population standard deviation. If 100 sample
means based on n observations each are computed, and 100 confidence
intervals are set up using the above formula, we expect that about 95 of the
100 intervals will include {i. If only one interval is set up based on one sample
of n, as is usually the case, we can state that we have 95 percent confidence
that this interval includes fi. If cr is unknown, the Student’s t distribution
(Appendix, Table B) is used, and confidence intervals are given by
Y ± tl-a/2
18.0 mi/gal (F) or less when tested. The critical region of a test statistic con-
sists of all values of the test statistic where the decision is made to reject
HQ. In the example above, the critical region for the test statistic Y is where
F < 18.0 mi/gal.
Since hypothesis testing is based on observed sample statistics computed
on n observations, the decision is always subject to possible errors. If the
hypothesis is really true and it is rejected by the sample, a type I error is
committed. The probability of a type I error is designated as a. If the hypoth-
esis is accepted when it is not true, that is, some alternative hypothesis is
true, a type II error has been made and its probability is designated as p.
These a and P error probabilities are often referred to as the risks of making
incorrect decisions, and one of the objectives in hypothesis testing is to de-
sign a test whose a and p risks are both small. In most such test procedures
Of is set at some predetermined level, and the decision rule is then formulated
in such a way as to minimize the other risk, p. In quality control work, a
is the producer’s risk and p the consumer’s risk.
To review hypothesis testing, a series of steps can be taken that will
apply to most types of hypotheses and test statistics. To help clarify these
steps and to illustrate the procedure, a simple example is given parallel to
the steps.
-z
-1.645 0
%
- 19.5
-1.645
2/V25
or Y^ = 18.8, and the decision rule can be expressed as: Reject HQ if Y <
18.8. Here Y = 18.9 and the hypothesis is not rejected.
The procedure outlined may be used to test many different hypotheses.
The nature of the problem will indicate what test statistic is to be used, and
proper tables can be found to set up the required critical region. Well known
are tests such as those on a single mean, two means with various assumptions
about the corresponding variances, one variance, and two variances. These
tests are reviewed in Sections 2.6-2.8.
18.8 - 18.0
Z = 2.00
04
and the P error is only 0.0228.
If P is plotted for various values of fi, the resulting curve is called the
operating characteristic (or OC) curve for this test. Table 2.1 summarizes
the computations for several assumed values of p.
2.5 How large a sample? 23
18.8 — p
p 2/V25 P 1 - P
* (approximately = a)
j3 (probability
The question of how large a sample to take from a population for making
a test is one often asked of a statistician. This question can be answered
provided the experimenter can answer each of the following questions.
24 Review of statistical inference
^ Reject * Accept ►
1. How large a shift (from 19.5 to 19.0) in a parameter do you wish to detect?
2. How much variability is present in the population? (Based on past expe-
rience, (7 = 2 mi/gal.)
3. What size risks are you willing to take? (a = 0.05 and p = 0.10.)
a equation:
F - 19.5
—— = — 1.645 (based on a = 0.05)
llsfn
„ .
P equation:
7 - 19.0
—— = +1.282 (based on p = 0.10)
2/y/n
Subtracting the second equation from the first and multiplying both sides
of each equation by 2/y/n gives
-0.5 - -2.921
5.854
11.71
0.5
19.22 mi/gal
The decision rule is then: Choose a random sample of 137 carburetors, and
if the mean mi/gal of these is less than 19.22, reject HQ ; otherwise, accept HQ.
Excellent tables are available for determining n for several tests of
hypotheses, such as those in Davies [9, pp. 606-615] or Owen [20, pp. 19,
23, 36, 41, ff.].
(n - l)s^ 2 /
2 < (^ < 2
Xa/2 ll-01.12
using the chi-square table with {n — 1) df to give 100(1 — a) percent con-
fidence limits.
As an example consider that one might wish to determine whether a
process standard deviation is greater than a specified value of 5 ounces. If
a sample of ten weights is found to have an average of 35.2 ounces and a
standard deviation of 7 ounces, what can we conclude about the standard
deviation of this process?
We hypothesize HQ: G^ = 25 oz^ and take as alternative H^ : G^ >
25 oz^. Selecting a = 0.05 and n = 10 gives n — 1 = 9 df. One should
reject HQ if
> ZL5(9 df) = 16.919
so we reject the hypothesis and conclude that our process standard deviation
is greater than 5 ounces.
We might wish to set 90 percent confidence limits on (or a) based on
our observed variance of 49:
9(49) 9(49)
< G^
Zo.05 Xo.95
9(49) ^ ^ 9(49)
16.919 “ “ 3.325
26.07 <<j^< 132.63
or
5.1 < cr < 11.5 ounces
The critical region is set at F > ^0.95 in this example (a == 0.05) with
the degrees of freedom dependent on the size of the two samples.
2.7 Application to tests on means 27
= 156
and the second are
^2 = 100
then
156
F - 1.56
100
and the critical region is F > 3.29 for 7 and 9 df. Hence the hypothesis is not
rejected.
In this example only a one-sided test has been considered as one is
usually concerned with showing that one variance is greater or less than
another. It is customary then to place the larger observed variance in the
numerator of F and call it s^.
Z is A(0, 1)
Reject HQ if |Z S Zi-ai2 [orZ > Zi_„ orZ < ZJ.
slyjn
70.8 - 75
t = = -1.39
7.4/V6
Since \t\ = 1.39 < 2.57, one cannot reject HQ. We might have used our
data to set 95 percent confidence limits on fi. From Section 2.2 100(1 — a)
percent confidence limits are
^ + h-aiisl-Jn
70.8 + 2.51(1 A) iS
70.8 + 7.8
HQ: = fi2
Y, - Y2
t =
l{n^ - l)si + (772 - l)s
77i + 772 — 2 77i ^ 772^
with 77^ + 772 — 2 df. One rejects if |t| > fi_<^/2 [or t > fi_J.
statistic
and check its significance using a special formula for its degrees of
freedom:
^0- I^D ~~ 0
HI: fij) ^ 0 (/rj5 > 0)
n differences, oc
Test statistic is
d
t = with n — 1 df
Some examples illustrating the foregoing tests are given in the following.
Example 1 If two samples randomly selected from two independent normal
populations give
X . ^2 = 4, F = 12.0,
2 si = 2.0, =9, Fi = 16.0 s\ = 5.0
^0- ~ 1^2
Hl'lll > fil
(jj = aj (note F = 2.5 < Tg,3(0.975))
30 Review of statistical inference
Reject if
^ ^ ^0.95,11 df ~ 1.V96
16.0 - 12.0
t = = 3.25
8(5) + 3(2) 1 1
11 9^4
Hence reject HQ and claim is greater than /i2-
Example 2 A group of ten students was pretested before instruction and
posttested after six weeks of instruction with the following achievement
scores.
2^*5
Is there evidence of an improvement in achievement over this six-week
period?
^0- I^D ~ 0
HI: p,j) > 0
Above
d = 2.5, = 2.22
t = —= 3.56
2.22/yio
Since t is >^9,0.95 1.833, we reject HQ and conclude that there has
been an improvement.
test statistic is
P - Po
^ ^ 10/25 - 0.20 ^
V(0.2)(0.8)/25
SOwe reject HQ and conclude that our class does have an excessive
number of left-handers.
^iPl + ^2P2
Hi -h ^2
^o-Pi = Pi
^1- Pi > Pi where p^ is proportion on Monday,
a = 0.05, = 100, ^2 = 200.
32 Review of statistical inference
PROBLEMS
2.1 For the following data on tensile strength in psi (pounds per square inch), deter-
mine the mean and variance of this sample:
Tensile Strength
(psi) Frequency
18,461 2
18,466 12
18,471 15
18,476 10
18,481 8
18,486 3
Total 50
2.2 Using the results of Problem 2.1, test the hypothesis that the mean tensile strength
of the population sampled is 18,470 psi (assume that = 40, a = 0.05).
2.3 Plot the operating characteristic curve for Problem 2.2.
2.4 Determine how large a sample would be needed to detect a 10-psi increase in the
mean tensile strength of Problem 2.2 if a = 0.05, f] = 0.02, and = 40.
2.5 Repeat Problem 2.4 for the detection of a 10-psi shift in the mean in either direction.
2.6 For a sample mean of 124 g (grams) based on 16 observations from a population
whose variance is 25 g^ set up 90 percent confidence limits on the mean of the
population.
2.7 For a sample variance of 62 based on 12 observations test the hypothesis that the
population variance is 40. Use a one-sided test and a 1 percent significance level.
2.8 For Problem 2.7 set up 90 percent confidence limits (two-sided) on
2.9 Two samples are taken, one from each of two machines. For this process
Hi =8, L = 42.2 g, sf = 10 g^
^2 = 15, F2 = 44.5 g, si = 18 g^
2.11 Reflection light box readings before and after dichromating the interior of a metal
cone were
Test
Number 1 2 3 4 5 6 7 8
2. If the process now shifts to a 0.10 fraction defective, find the probability of an
error of the second kind for this alternative.
3. Without repeating the work, would you expect the j? error to be the same if the
process shifted to 0.30? Explain.
2.13 Given the following sample data on a random variable, 7: 12, 8, 14, 20, 26, 26, 20,
21, 18, 24, 30, 21, 18, 16, 10, 20. Assuming cr = 7, and a = 0.05, test each of the
following hypotheses:
1. HQ: P = 12
H^ : p > 12
2. Ho : p = 16
H ^; p 7^ 16
3. HQ] P = 18
/flip > 18
2.14 For test 1 in Problem 2.13 evaluate the power of the test with respect to several
alternative hypotheses from p = 13 to p = 24. Plot the power curve.
2.15 Repeat the results of Problem 2.14 with cc — 0.01. Plot on the same graph and
compare the two power curves.
2.16 Given the following data on the current flow in amperes through a cereal forming
machine: 8.2, 8.3, 8.2, 8.6, 8.0, 8.4, 8.5, 7.3. Test the hypothesis of the standard
deviation a = 0.35 ampere versus the alternative that a > 0.35 ampere.
2.17 For the data in Problem 2.16 test the hypothesis that the true mean current flow
is p = 7.0.
2.18 The percent moisture content in a puffed cereal where samples are taken from two
different “guns” showed
Test the hypothesis of equal variances and equal means. Use any assumptions you
believe appropriate.
34 Review of statistical inference
2.19 Pretest data for experimental and control groups on course content in a special
vocational-industrial course indicated:
Experimental: = 9.333, = 4.945, = 12
Control: Y2 = 8.375, S2 = 1.187,^2 = 8
Make any statistical tests that the data might suggest. Comment on the results.
2.20 The WISC performance scores of ten boys and eight girls were recorded as follows:
Boys: 101, 86, 72, 129, 99, 118, 104, 125, 90, 107
Girls: 97, 107, 94, 101, 90, 108, 108, 86
Test whether or not there is a significant difference in the means of boys and girls.
Take a = 0.10.
2.21 An occupational therapist conducted a study to evaluate the relative merits of two
prosthetic devices designed to facilitate manual dexterity. The therapist assigned
21 patients with identical handicaps to wear one or the other of the two devices
while performing a certain task. Eleven patients wore device A and ten wore device
B. The researcher recorded the time each patient required to perform a certain
task, with the following results:
— 65 seconds = 81
yg = 75 seconds s| = 64
Do these data provide sufficient evidence to indicate that device A is more effective
than device B1
2.22 An anthropologist believes that the proportion of individuals in two populations
with double occipital hair whorls is the same. To see whether there is any reason
to doubt this hypothesis, the anthropologist takes independent random samples
from each of the two populations and determines the number in each sample with
this characteristic. Results showed:
1 100 23
2 120 32
Scores
Wife 33 57 32 54 52 34 60 40 59 39
Husband 44 60 55 68 40 48 57 49 47 52
Problems 35
2.24 Measurements were made at each of two plants on the tensile strength of samples
of the same type of steel. The data are given below in thousands of psi. Can it be
stated that the steels made in the two plants have the same mean tensile strength?
Plant 1 Plant 2
207 219
195 228
209 222
218 197
196 225
217 184
218
211
2.25 The standard medical treatment for a certain ailment has proved to be about
85 percent effective. One hundred patients were given an alternative treatment
with the result that 78 showed marked improvement. Is the new treatment actually
inferior or could the difference noted be reasonably attributed to chance?
3
Single-Factor Experiments
with No Restrictions
on Randomization
3.1 INTRODUCTION
In this and several subsequent chapters single-factor experiments will be
considered. In this chapter no restrictions will be placed on the randomiza-
tion so that the design will be completely randomized. Many of the tech-
niques of analysis for a completely randomized single-factor experiment can
be applied with little alteration to more complex experiments.
For example, the single factor could be steel manufacturers, where the
main interest of an analyst centers on the effect of several different manu-
facturers on the hardness of steel purchased from them. It could be tem-
perature in an instance where the experimenter is concerned about the effect
of temperature on penicillin yield. Whenever only one factor is varied,
whether the levels be quantitative or qualitative, fixed or random, the experi-
ment is referred to as a single-factor experiment, and the symbol TJ will be
used to indicate the effect of the jth level of the factor. TJ suggests that the
general factor may be thought of as a “treatment” effect.
If the order of experimentation applied to the several levels of the
factor is completely random, so that any material to which the treatments
might be applied is considered approximately homogeneous, the design is
called a completely randomized design. The number of observations for each
level of the treatment or factor will be determined from cost considerations
and the power of the test. The model then becomes
fj = p Tj + Sfj
where Y^j represents the ith observation (/ = 1, 2,..., nf on the jth treat-
ment (j = 1, 2,..., /c levels). For example, T23 represents the second ob-
servation using level 3 of the factor, p is a common effect for the whole
experiment, TJ represents the effect of the jth treatment, and represents
the random error present in the ith observation on the jth treatment.
The error term 8,^ is usually considered a normally and independently
distributed (NID) random effect whose mean value is zero and whose
36
3.1 Introduction 37
variance is the same for all treatments or levels. This is expressed as: 8j/s
are NID (0, where is the common variance within all treatments.
is always a fixed parameter, and T2,..., TJ, .. ., are considered to
be fixed parameters if the levels of treatments are fixed. It is also assumed
that
k
J=1
If the k levels of treatments are chosen at random, the T/S are assumed
NID (0, (T^). Whether the levels are fixed or random depends upon how
these levels are chosen in a given experiment.
The analysis of a single factor completely randomized experiment
usually consists of a one-way analysis of variance test where the hypothesis
HQ : TJ = 0 for all j, is tested. If this hypothesis is true, then no treatment
effects exist and each observation Yij is made up of its population mean ji
and a random error Sij. After an analysis of variance (ANOVA), many other
tests may be made, and some of these will be shown in the example that
follows.
Yij = Y Tj &ij
with i = 1, 2,.. ., 4 and j = 1, 2,.. ., 4 since there are four samples for
each of four treatments (fabrics). The data are shown in Table 3.1.
Fabric
A B C D
Source of Variation df SS MS
Totals 15 0.7639
0.1734
8.54
0.0203
which is significant at the 1 percent significance level since Appendix Table D
shows 5.95 as the 1 percent F {FQ gg) for 3 and 12 degrees of freedom. We
can then reject the hypothesis and claim that there is a considerable differ-
ence in average wear resistance among the four fabrics.
Treatment
1 2 • • • j • • • k
hi Yn Yu
Here the use of the “dot notation” indicates a summing over all observa-
tions in the population. Since each observation could be returned to the
population and measured, there could be an infinite number of observations
3.2 Analysis of variance rationale 39
This last expression is seen to be an identity true for all values of Y^j. Ex-
pressed another way,
Yij - fi = {fij - fi) Y {Yij - (3.1)
Since these means are unknown, random samples are drawn from each
population and estimates can be made of the treatment means and the grand
mean. If rij observations are taken for each treatment where the numbers
need not be equal, a sample layout would be as shown in Table 3.4.
Treatment
1 2 J k
Ell EI2
... Y^j ••• Eu
E21 E22 ... Y2J E2.
;
E«.2 Ynfji
r. = I I J = 1 I= 1
Yii = Z
j= 1
Tj
and
N= Z nj
40 Single-factor experiments with no restrictions on randomization
Y.= i njYJN
J=1
This equation states that the deviation of any observation from the grand
mean can be broken into two parts: the deviation of the observation from its
own treatment mean plus the deviation of the treatment mean from the
grand mean.
If both sides of Equation (3.2) are squared and then added over both i
and j, we have
i I
j=l 1= I
(Y, - = i I
j=l 1=1
(Fj - F.)^
j=l 1=1
(Yu - Yjf
y y (F, -
/= 1 ^ = 1
- F,) = y (y, - yj
7= 1
i
i= 1
(Yij - F,)
The term in brackets is seen to equal zero, as the sum of the deviations about
the mean within a given treatment equals zero. Hence
(3.4)
1
^
i= 1
= I I 1 i= 1
(^J - F..)' + Z I
7=1 i =1
(Yij- Y.jf
/0/]no\A9.
This may be referred to as “the fundamental equation of analysis of variance,”
and it expresses the idea that the total surn of squares of deviations from the
grand mean is equal to the sum of squares of deviations between treatment
means and the grand mean plus the sum of squares of deviations within
treatments. In Chapter 2 an unbiased estimate of population variance was
determined by dividing the sum of squares (1^ — Y)^ by the corre-
sponding number of degrees of freedom, — 1. If the hypothesis being
tested in analysis of variance is true, namely, that Zj = 0 for all j, or that
there is no treatment effect, then i ^ i^.2 ^ ‘ ‘ ‘= /kfc
there is in the model is the population mean fi and random error Then,
any one of the three terms in Equation (3.4) may be used to give an unbiased
estimate of this common population variance. For example, dividing the
3.2 Analysis of variance rationale 41
y - 1) = N - fc
j
which will give another estimate of the population variance. Still another
estimate can be made by first estimating the variance (Ty between means
drawn from a common population with cry = njGy An unbiased estimate
for (jy is given by “ Y_)^/{k — 1) so that an unbiased estimate of
= i MYj - Y.f/{k - 1)
j=i
which is found by summing the first term on the right-hand side of Equa-
tion (3.4) over i and dividing by /c — 1 df. Thus there are three unbiased
estimates of possible from the data in a one-way ANOVA if the hypothesis
is true. Now all three are not independent since the sum of squares is additive
in Equation (3.4). However, it can be shown that if each of the terms (sums
of squares) on the right of Equation (3.4) is divided by its proper degrees of
freedom, it will yield two independent chi-square distributed unbiased esti-
mates of cr^ when HQ is true. If two such independent unbiased estimates of
the same variance are compared, their ratio can be shown to be distributed
as F with k — 1, A — /c df. If, then, HQ is true, the test of the hypothesis
can be made using a critical region of the F distribution with the observed
F at k — 1 and N — k dt given by
. lUinjiYj - ff/{k- 1}
Z5=i (Yj - Yjf/iN - k)
The critical region is usually taken as the upper tail of the F distribution,
rejecting HQ if T > F^_^ where <x is the area above In this F ratio the
sum of squares between treatments is always put into the numerator, and
then a significant F will indicate that the differences between means has
something in it besides the estimate of variance. It probably indicates that
there is a real difference in treatment means (/i i, 2^ • • •) that HQ
should be rejected. These unbiased estimates of population variance, sums
of squares divided by df, are also referred to as mean squares.
The actual computing of sums of squares indicated in Equation (3.5) is
much easier if they are first expanded and rewritten in terms of treatment
totals. These computing formulas are given in Table 3.5. Applying the
42 Single-factor experiments with no restrictions on randomization
Source df SS MS
T^j
Within treatments
or error Sij
N - k it
j= l i=l
SSerror/CA" “
fc Tij h '7^2
=
m " tij
k tij
Totals N - 1 EE
j=l i=l
(F« - F/
k nj rj.2
J=1 i=l
formulas in Table 3.5 to the problem data in Table 3.1, Table 3.6 shows
treatment totals T /s and the sums of squares of the YJ/s for each treatment
along with the results added across all treatments.
Fabric
A B C D
The sums of squares can be computed quite easily from this table. The
total sum of squares states, “Square each observation, add over all observa-
tions, and subtract the correction term.” The last is the grand total squared
3.2 Analysis of variance rationale 43
+ + +CREATE PROB7
1. 000 = 77509,113
2. 000 = COMMON (SPSS)
3. 000 = SPSS
4. 000 = #EOR
5. 000 = RUN NAME STAT 502 PROB7
6. 000 = VARIABLE LIST WE,FAB
7. 000 INPUT FORMAT (F4.0, IX, FLO)
8. 000 = N OF CASES 16
9. 000 = ONEWAY WE BY FAB (1,4)/
10. 000 = CONTRAST =10 0 -1/
11. 000 = CONTRAST = 01-10/
12. 000 CONTRAST =1-1-11/
13. 000 = RANGES = SNK
14. 000 STATISTICS AEL
15. 000 = READ INPUT DATA
16. 000 = 1.93 1
17. 000 = 2.38 1
18. 000 = 2.20 1
19. 000 = 2.25 1
20. 000 = 2.55 2
21. 000 = 2.72 2
22. 000 = 2.75 2
23. 000 = 2.70 2
24. 000 2.40 3
25. 000 = 2.68 3
26. 000 = 2.31 3
27. 000 2.28 3
28. 000 = 2.33 4
29. 000 = 2.40 4
30. 000 = 2.28 4
31. 000 = 2.25 4
32. 000 = FINISH
33. 000 = #s
CQ [JH
z ON CN o
oq
T-H
o <N z <!
W
r4 (N (N (N
osi
cu
O
o
o
u
o
HoHo
HoH
[JU ON m m r-
OO oo r- o
OO m (N 1—H
oq in <N
m i-H c4 r4 r4
O ON
l—H
H m
< in
eej od
[JH P o o o o
2.7500
o o o o
l-H oo m oo o
rn r- NO
Z ni (N (N (N
00 <
W
cd
<
p m
a m o P o o O
1.9300
r- <N o
oo T-H o o O o
m m oo m
W Z HH ON m rsi <N
U < z
HH
<N (N r4
Z W
<
»-H s
osi Q
<
^ 00
OP OP
< NO in oo
O (N
Q OP ON ON m
O ^ Z OP o o o O
< w
“
00 S. oo o^ H
00
oo o m m
(N
m (N r--
00
00
<
1^
o
Q z
oP o
DH < < H
HH
(N m NO
00
ON ON <N in
Q < OO OO OO NO
p
Xi 00 rn Z t-H o o
< >
rn "a H w
oo Q
a
'B. P ro <N m oj
X
Z o o in o
2.4006
S Q w
< o o r- in
X c ON oo 1—H H
PQ o w NO rn
U4
u < (N (N r4 (N
-t-j
i2 cij
< >-' Q H
> CQ
OH z
16
O HJ w =S p
o
z u
pcS
Ui
o
o
< w
u
HJ
P
pin
TOTAL
oc O o\
PQ
00 ro ^ <N m rj-
fO <
H-: P QH OH PH PH
2 pi^ 2
2
O OP OP OP OP
SQ < OP
H O O O O O
H >
AS.
46 Single-factor experiments with no restrictions on randomization
This table shows the mean and standard deviation of each group—
or treatment level—1 through 4 and its corresponding standard error (sy =
SyZ-s/n) and 95 percent confidence limits on its true group mean.
c = y c T
j=l
where
E Cjn, = 0
7=1
In dealing with several contrasts it is highly desirable to have the
contrasts independent of each other. When the contrasts are independent—
one having no projection on any of the others—they are said to be orthogonal
contrasts, and independent tests of hypotheses can be made by comparing
48 Single-factor experiments with no restrictions on randomization
the mean square of each such contrast with the mean square of the error term
in the experiment. Each contrast carries one degree of freedom.
It may be recalled that when two straight lines, say
CiiX + ^217 =
^12^ + ^22^ — ^2
are perpendicular, or orthogonal, to each other, the slope of one line is the
negative reciprocal of the slope of the other line. Changing the two lines
above into slope-intercept form.
Y = -C11/C21X + 61/C21
7 = ~ ^12/^22^ + ^2/^22
To be orthogonal,
~^lll^2l — ~{~ ^22/^12)
or
^11^12 — “^ 21^^22
or
^11^12 + <^21^22 ~ 0
The sum of the products of the corresponding coefficients (on X and
on 7) must therefore add to zero. This idea may be extended to a general
definition.
Two contrasts and are said to be orthogonal contrasts, provided
k
= 0
7=1
SS
= - 7.4
C2 = 7.2 - 7.3
T.i T.2 T3 T4
Cl +1 0 0 -1
C2 0 +1 -1 0
C3 +1 -1 -1 +1
It can be seen from Table 3.10 that the sum of coefficients adds up
to zero for each contrast, and that the sum of products of coefficients of
each pair of contrasts is also zero.
For the example given in Section 3.2, using the data in Table 3.6, the
contrasts are
(-0.50)^
0.0312
4(2)
(1.05)2
0.1378
4(2)
(-2.37)2
0.3511
^(4F
Total = 0.5201
Note that these three orthogonal sums of squares add up to the treatment
sums of squares. Each contrast has 1 df, so one may test hypotheses implied
by the contrasts as follows:
0.0312/1
= ^^4 or -^1,12 1.54
0.0203
t
0.1378/1
Ho,- ^2 = ^^3 or /^2 = 1^3 ^1,12 6.76
0.0203
0.3511/1
Ho,-Ti + T4. = '^2 + '^3 ■^1,12 17.30
0.0203
or
+4 — 1^2 + 3
50 Single-factor experiments with no restrictions on randomization
Since the F for 1 and 12 df is 4.75 at the 5 percent significance level and 9.33
at the 1 percent significance level, one can reject HQ^ at the 5 percent level
and conclude that the mean wear resistance of fabric 2 (or B) differs from that
of fabric 3 (or C). At the 1 percent level one rejects and concludes that the
mean wear resistances of fabrics 1 and 4 {A and D) differ from that of 2 and 3
{B and C).
If a computer program is used to run the ANOVA, it can be programmed
to compute any desired contrasts. In Table 3.7 lines 10, 11, and 12 indicate
the three contrasts desired. The printout is as shown in Table 3.11.
At first glance these results seem different from the calculated values
above; however, the “value” of the contrast given in Table 3.11 is based on
means instead of totals, so multiplying each by rij = 4 yields the contrasts
of —0.5, 1.05, and —2.37. Then the tests are run as t tests instead of F tests.
In the case of 1 df it can be shown that tjf = ^f. Here, for example, for the
second contrast t = 2.6, which squares into 6.76 as seen above.
As this method of orthogonal contrasts is used quite often in experi-
mental design work, the definitions and formulas are given below for the
case of unequal numbers of observations per treatment.
Cjn is a contrast if
k
Z "jCjm = 0
J=1
k
I = 0
j= 1
3.4 Tests on means 51
If the decision on what comparisons to make is withheld until after the data
are examined, comparisons may still be made, but the cc level is altered
because such decisions are not taken at random but are based on observed
results. Several methods have been introduced to handle such situations
[17], but only the |^ewimn-I<reuls] range test [15] and Scheffe’s test [21]
are described here.
Range Test After the data have been compiled the following steps are taken.
1. Arrange the k means in order from low to high.
2. Enter the ANOVA table and take the error mean square with its degrees
of freedom.
3. Obtain the standard error of the mean for each treatment
where the error mean square is the one used as the denominator in the
F test on means Yj.
4. Enter a Studentized range table (Appendix, Table E) of significant ranges
at the a level desired, using ^2 = degrees of freedom for error mean
square and p = 2, 3,. .., /c, and list these k — 1 ranges.
5. Multiply these ranges by Sy . to form a group of /c — 1 least significant
ranges.
6. Test the observed ranges between means, beginning with largest versus
smallest, which is compared with the least significant range for p = k;
then test largest versus second smallest with the least significant range
for p = k — 1; and so on. Continue this for second largest versus
sm.allest, and so forth, until all k{k — l)/2 possible pairs have been tested.
The sole exception to this rule is that no difference between two means
can be declared significant if the two means concerned are both contained
in a subset with a nonsignificant range.
To see how this works, consider the data of Example 3.1 as given in
Table 3.6. Following the steps given above:
1. /c = 4 means are 2.19 2.32 2.42 2.68
for treatments: A D C B
52 Single-factor experiments with no restrictions on randomization
2. From Table 3.2 the error mean square is 0.0203 with 12 df.
3. Standard error of a mean is
/0.0203
0.0712
p: 2 3 4
Ranges: 3.08 3.77 4.20
From the results in step 6 one sees that B differs significantly from
A, D, and C, but A, Z), and C do not differ significantly from each other.
These results are shown in Figure 3.2, where any means underscored by
the same line are not significantly different.
Treatment: D C B
Al^
i
Schejfe’s Test Since the Newman-Keuls test is restricted to comparing pairs
of means and it is often desirable to examine other contrasts which represent
combinations of treatments, many experimenters prefer a test devised by
Scheffe [21]. Scheffe’s method uses the concept of contrasts presented earlier
but the contrasts need not be orthogonal. In fact, any and all conceivable
contrasts may be tested for significance. Since comparing means in pairs is a
special case of contrasts, Scheffe’s scheme is more general than the Newman-
Keuls. However, since the Scheffe method must be valid for such a large set
3.4 Tests on means 53
Applying this technique to the coded data of Tables 3.1, 3.2, and 3.6,
proceed as follows.
(Note that these are not orthogonal. The first compares treatment A
with B and the second compares the mean of treatment A with the mean
of the other three treatments.)
2. Since a = 0.05, /c - 1 = 3, iV - /c = 12, F = 3.49.
3. A = V3(3.49) = 3.24.
4. Sc, = V0.0203[4(l)^ + 4(-l)^] = 0.40 and Asc, = 1.30.
Sc, = V0.0203 [4(3)^ + 4( - 1 )^ + 4( -1)^ + 4( - 1 )^] = 0.99 and dsc, = 3.21,
5. Since |Ci| = 1.96 is > 1.30, this contrast is significant.
Since C2 = 3.37 is >3.21, this contrast is also significant.
SUBSET 1
GROUP GRP 1 GRP 4 GRP 3
MEAN 2.1900 2.3150 2.4175
SUBSET 2
GROUP GRP 2
MEAN 2.6800
where the mean square used to test treatment mean square is the error mean
square in a one-way ANOVA but may be a different mean square in more
complex analyses. The degrees of freedom used with the Student’s t statistic
are the degrees of freedom that correspond to the mean square used for
testing the treatment mean.
In Example 3.1,95 percent confidence limits on the mean wear resistance
for fabric B, /i2 would be
except that in this experiment the k levels of the treatment (batches) are
chosen at random, rather than being fixed levels. The data are shown in
Table 3.13.
Batch
1 2 3 4 5
74 68 75 72 79
76 71 77 74 81
75 72 77 73 79
Source df SS MS EMS
Totals 14 165.73
n(k — 1) . ,
EMS,„„, = (7? + o-£
Since the coefficient of is always less than one, this EMS will be less than the
total variance given above as aj.
SS.,eatment = I n{Yj - Y_ ^
J=1
From model Equation (3.7)
n
nU HTi A
= -r
n + n ~+ Z i= 1
71
= + Zi= 1
(3.8)
Also,
k ti k fj
Y. = Z Z Ey/n/c = Z Z (/i + Tj + Sij}/nk
j=l i = 1 j i
nkp. V ; I VV / )
nk + «Zj + Zj ZI
k k n
= /^ + Z + ZZ
J I
(3.9)
Squaring gives
n k
(Y.j - r f = (Tj - z ^ ^
(+ cross products)
58 Single-factor experiments with no restrictions on randomization
= »! Z bj - I
n k
n
+ ri Z Z e.j “ I Z + « Z (cross product)
^ j \ i i j J j
EiSS treatment ) = nE IJ \
- IJ
1 k / n n k
+ -E
n Z Z s.j - Z Z s.jA
J \ I ' J
as it can be shown that the expected value of the cross-product term equals
zero.
If the treatment levels are fixed,
j=i
Z = Z (t'.j - /i) = 0
j=i
T(SS treatment
7=1
since errors are random and Z7=i ^ constant. The T(MStreatment) =
T [^SStreatment/(^
k
n{k - 1)
£(MS treatment ) = " Z 1) + (J,
L 7=1
-
n{k - 1)
If, on the other hand, treatment levels are random and their variance
IS (T:
^(MSt.eatment) ==
(jg + n Z - 1)
L 7
or
(Jg + UG^
3.7 Summary 59
SSerror = X I(
7=1 «=1
Subtracting Equation (3.8) from Equation (3.7) gives
^ij ^ij
i= 1
X ^ij/^
ij 'n
j I j I \ I
= 1E
J
I -I n
= Z - 1)'^^
j
= k{n -
and
£(MS„,J = E[SS,„Jk{n - 1)] = af
as shown in Table 3.14.
3.7 SUMMARY
In this chapter consideration has been given to
Many texts [4], [9] discuss these assumptions and what may be done if
they are not met in practice.
Several studies have shown that lack of normality in the dependent
variable Y does not seriously affect the analysis when the number of observa-
tions per treatment is the same for all treatments. To check for homogeneous
variances within treatments, several tests have been suggested. One quick
check is to examine the ranges of the observations within each treatment.
For Example 3.1 these are 0.45, 0.20, 0.40, and 0.15. If the average range
R is multiplied by a factor D4, which is found in most quality control texts,
and all ranges are less than D4_R, it is quite safe to assume homogeneous
variances. Here R = 0.30 and D^^R = 0.68 based on samples of four. Since
all four ranges are well below 0.68, homogeneity of variance may be reason-
ably assured. Values of for a few selected sample sizes are given in
Table 3.15.
U n = 2 3 4 5 6 7 8 9 10
D4 = 3.267 2.575 2.282 2.115 2.004 1.924 1.864 1.816 1.777
With some measured variables it may be that the means of the treat-
ments and the variances within the treatments are related in some way. For
example, a Poisson variable has its mean and variance equal. It can be
shown mathematically that a necessary and sufficient condition for normality
of a variable is that means and variances drawn from that population are
independent of each other. If it is known or observed that sample means and
sample variances are not independent but are related to each other, it may
be possible to “break” this correlation by a suitable transformation of the
original variable Y. Table 3.17 gives several suggested transformations to
try if certain specified relationships hold.
Problems 61
If is proportional to Transform Lj to
PROBLEMS
Factor A
Level
1 2 3 4 5
A
Measurement 8 4 1 4 10
6 -2 2 6 8
J 0 0 5 7
5 -2 -1 5 4
8 3 -3 4 9
The cathode warm-up time in seconds was determined for three different tube
types using eight observations on each type of tube. The order of experimentation
was completely randomized. The results were
Tube Type
ABC
Warm-up time 19 20 20 40 16 19
(seconds) 23 20 20 24 15 17
26 18 32 22 18 19
18 35 27 18 26 18
or ;
Do an analysis of variance on these data and test the hypothesis that the three
tube types require the same average warm-up time.
■n
P ^ /'-C
0-^ rr^:Xyn^/^^^ i w^
For Problem 3.2 set up orthogonal contrasts between the tube types and test your
contrasts for significance.
3.4 )Set up 95 percent confidence limits for the average warm-up time for tube type
C in Problem 3.2.
3.5 Use the Newman-Keuls range method to test for differences between tube types.
3.6 The following data are on the pressure in a torsion spring for several settings of
the angle between the legs of the spring in a free position.
67 71 75 79 83
Pressure (psi) 83 84 86 89 90
85 85 87 90 92
85 87 90
86 87 91
86 88
87 88
88
88
88
89
90
Temperature (°C)
0 25 50 75 100
Activated life 55 60 70 72 65
(seconds) 55 61 72 72 66
57 60 73 72 60
54 60 68 70 64
54 60 77 68 65
56 60 77 69 65
3.14 A highway research engineer wishes to determine the effect of four types of sub-
grade soil on the moisture content in the top soil. He takes five samples of each
type of subgrade soil and the total sum of squares is computed as 280, whereas
the sum of squares among the four types of subgrade soil is 120.
1. Set up an analysis-of-variance table for these results.
2. Set up a mathematical model to describe this problem, define each term in
the model, and state the assumptions made on each term.
3. Set up a test of the hypothesis that the four types of subgrade soil have the
same effect on moisture content in the top soil.
4. Set up a set of orthogonal contrasts for this problem.
5. Explain briefly how to set up a test on means after the analysis of variance
for these data.
6. Set up an expression for 90 percent confidence limits on the mean of type 2
subgrade soil. Insert all numerical values that are known.
3.1^ Data collected on the effect of four fixed types of television tube coating on the
conductivity of the tubes is given as
Coating
II III IV
56 64 45 42
55 61 46 39
62 50 45 45
59 55 39 43
60 56 43 41
Do an analysis of variance on these data and test the hypothesis that the four
coatings yield the same average conductivity.
3.16 For Problem 3.15 set up orthogonal contrasts between tube coatings and test
them for significance.
3.17 jUse a Newman-Keuls range test on the data of Problem 3.15 and discuss any
significant results that you find.
64 Single-factor experiments with no restrictions on randomization
3.18 Set up contrasts and test using Scheflfe’s method to answer the following questions:
Do the means of coating I and II differ? The means of II and IV? The mean of
I and II versus the mean of III and IV ?
3.19 Explain why confidence limits on a treatment mean as given in Table 3.9 might
differ from the method suggested in Section 3.5. Which method would you recom-
mend under what conditions?
y 3.2(y To determine the effect of several teaching methods on student achievement, 30
students were assigned to five treatment groups with six students per group. The
treatments given for the semester are as follows:
Treatment Description
1 Current textbook
2 Textbook A with teacher
3 Textbook A with machine
4 Textbook B with teacher
5 Textbook B with machine
Treatment
1 2 3 4 5
Totals 120 600 720 240 420
Source df SS MS F F 0.95
Total 465
3.24 Three fertilizers are tried on 27 plots of land in a random fashion such that each
fertilizer is applied to nine plots. The total yield for each fertilizer type is given by
Type
1 2 3
Tj 240 320 180
Source df SS MS
Total 26 2062
a. Set up one set of orthogonal contrasts that might be used on these data.
b. For your first contrast, determine the sum of squares due to this contrast.
c. For your second contrast, find its standard error.
d. If one wished to compare types 1 and 2 and also the average of 1 and 3 versus
2, which method would you recommend and why?
3.25 Birth weights in pounds of five Poland China pigs were recorded from each of six
randomly chosen litters. The ANOVA printout showed:
Source df SS MS
Determine what percentage of the total variance in this experiment can be attrib-
uted to litter to litter differences.
4
Single-Factor Experiments —
Randomized Block
and Latin Square Designs
4.1 INTRODUCTION
Consider the problem of determining whether or not different brands of tires
exhibit different amounts of tread loss after 20,000 miles of driving. A fleet
manager wishes to consider four brands that are available and make some
decision about which brand might show the least amount of tread wear after
20,000 miles. The brands to be considered are A, B, C, and D, and although
driving conditions might be simulated in the laboratory, he wants to try these
four brands under actual driving conditions. The variable to be measured is
the difference in maximum tread thickness on a tire between the time it is
mounted on the wheel of a car and after it has completed 20,000 miles on
this car. The measured variable Y^j is this difference in thickness in mils
(0.001 inch), and the only factor of interest is brands, say, Zj where j = 1,2,
3, and 4.
Since the tires must be tried on cars and since some measure of error is
necessary, more than one tire of each brand must be used and a set of four
of each brand would seem quite practical. This means 16 tires, four each
of four different brands, and a reasonable experiment would involve at least
four cars. Designating the cars as I, II, III, and IV, one might put brand T’s
four tires on car I, brand B's on car II, and so on, with a design as shown
in Table 4.1.
Car
I II III IV
Brand A B C D
distribution A B C D
A B c D
A B c D
66
4.2 Randomized complete block design 67
Car
I II III IV
One look at this design shows its fallibility, since averages for brands
are also averages for cars. If the cars travel over different terrains, using
different drivers, any apparent brand differences are also car differences.
This design is called completely confounded, since we cannot distinguish
between brands and cars in the analysis.
A second attempt at design might be to try a completely randomized
design, as given in Chapter 3. Assigning the 16 tires to the four cars in a
completely random manner might give results as in Table 4.2. In Table 4.2
the loss in thickness is given for each of the 16 tires. The purpose of complete
randomization here is to average out any car differences that might affect
the results. The model would be
with
7 = 1, 2, 3, 4 i = 1, 2, 3, 4
Source df SS MS
Totals 15 80.94
A more careful examination of design 2 in Table 4.2 will reveal some glaring
disadvantages of the completely randomized design in this problem. One
thing to be noted is that brand A is never used on car III or brand B on
68 Single-factor experiments—randomized block and latin square designs
car I. Also, any variation within brand A may reflect variation between cars
I, II, and IV. Thus the random error may not be merely an experimental
error but may include variation between cars. Since the chief objective of
experimental design is to reduce the experimental error, a better design
might be one in which car variation is removed from error variation. Al-
though the completely randomized design averaged out the car effects, it
did not eliminate the variance among cars. A design that requires that each
brand be used once on each car is a randomized complete block design, given
in Table 4.4.
Car
I II III IV
In this design the order in which the four brands are placed on a car is
random and each car gets one tire of each brand. In this way better compari-
sons can be made between brands since they are all driven over approximately
the same terrain, and so on. This provides a more homogeneous environment
in which to test the four brands. In general, these groupings for homogeneity
are called blocks and randomization is now restricted within blocks. This
design also allows the car (block) variation to be independently assessed and
removed from the error term. The model for this design is
where pi now represents the block effect (car effect) in the example above.
The analysis of this model is a two-way analysis of variance, since the
block effect may now also be isolated. A slight rearrangement of the data in
Table 4.4 gives Table 4.5.
The total sum of squares is computed as in Chapter 3
4 4
SS total
j I
(193)2
= 2409 - = 80.94
16
4.2 Randomized complete block design 69
Brand
Car A B C D T,
I 17 14 12 13 56
II 14 14 12 11 51
III 13 13 10 11 47
IV 13 8 9 9 39
Tj 57 49 43 44 = 193
T.
in
i
823 625 469 492 izn
j '
= 2409
T
SS brands
J ^ N
Since the car (block) effect is similar to the brand effect but totaled across
the rows of Table 4.5, the car sum of squares is computed exactly like the
brand sum of squares, using row totals instead of column totals. Calling
the number of treatments (brands) in general /c, then
n '-pi
SS...
car = T ^
i^
=1k 77"
The error sum of squares is now the remainder after subtracting both brand
and car sum of squares from the total sum of squares
Source df SS MS EMS
Totals 15 80.94
scheme will help the observer in noting just how the randomization has been
restricted.
Since blocking an experiment is a very useful procedure to reduce the
experimental error, a second example may be helpful as it illustrates the
situation in which the same experimental units (animals, people, parts, etc.)
are measured before the experiment and then again at the conclusion of the
experimental treatment. These before-after, pretest-posttest designs can be
treated very well as randomized block designs where the experimental units
are the blocks. Interest is primarily in the effect of the treatment and the
block effects can be removed to reduce the experimental error. Some refer
to these designs as repeated-measures designs because data are repeated on
the same units a second time. The following example will illustrate the
procedure.
1 100 115
2 no 125
3 90 105
4 no 130
5 125 140
6 130 140
7 105 125
2 no 125 235
3 90 105 195
4 no 130 240
Source df SS MS F Probability
Totals 13 2985.71
== F + ft + (4-2)
or
Yij = M + (/^i. - M) + iji.j - tA + (Yij - Hi, - H.j + A<) (4-3)
where represents the true mean of block 1 The last term can be obtained
by subtracting the treatment and block deviations from the overall deviation
as follows:
[Yij - H) - {Hi. - H) - (H.j - H) = Yij - Hi. - H.J + H
Best estimates of the parameters in Equation (4.3) give the sample model
(after moving Y to the left of the equation):
i i
I= 1 7= 1
{Yu - ry = ii(Yu - Yf + tiifj-
i j i j
F.p
+ iiiYu- Yu - Yj+ Yf
I J
(nk — 1) — (n — 1) — (k — 1) = nk — n — k -h 1 = (n — l)(k — 1)
It can be shown that each sum of squares on the right of Equation (4.4)
when divided by its degrees of freedom provides mean squares that are
independently chi-square distributed, so that the ratio of any two of them
is distributed as F.
The sums of squares formulas given in Equation (4.4) are usually
expanded and rewritten to give formulas that are easier to apply. These are
shown in Table 4.9. They are the formulas applied to the data of Table 4.5
where nk = N.
Source df SS MS
n T' 2 T2
Between n — \ S^block/(^ 1)
blocks Pi nk
k ’’p 2 T2
Between k - 1
E — - SStreatment/(^ 1)
treatments TJ j nk
n k ^ 2
Error 8ij {n - m- 1)
Ei E -
j i=l ^
k
-E^
r 2
Li + L.
^ 2
SS„„,/(n - 1)(A: - 1)
J=l n nk
n k
Totals nk — \
Ei Ej nk
blocks and treatments are orthogonal, the block totals are added over all
treatments, and vice versa. If one or more observations are missing, the usual
procedure is to replace the value with one that makes the sum of the squares
of the errors a minimum.
In the tire brand test example suppose that the brand C tire on car III
blew out and was ruined before completing the 20,000 miles. The resulting
data after coding by subtracting 13 mils appear in Table 4.10, where y is
inserted in place of this missing value.
Brand
Car A B C D T,
I 4 1 -1 0 4
II 1 1 -1 -2 -1
III 0 0 y -2 .V - 2
IV 0 -5 -4 -4 -13
T., 5 -3 V — 6 -8 y - 12 = T
Now,
SS error = SS total SS treatment - SS block
Tl
+
n k nk
4.4 Missing values 75
16_y — 4y + 24 — 4y + S -y y — 12 = 0
9y = -20
20
y —2.2 or 10.8 mils
y
If this value is now used in the y position, the resulting ANOVA table is
as shown in Table 4.11. This is an approximate ANOVA as the sums of
squares are slightly biased. The resulting ANOVA is not too different from
before, but the degrees of freedom for the error term are reduced by one,
since there are only 15 actual observations, and y is determined from these
15 readings.
This procedure can be used on any reasonable number of missing values
by differentiating the error sum of squares partially with respect to each
such missing value and setting it equal to zero, giving as many equations as
unknown values to be solved for these missing values.
Source df SS MS
Totals 14 78.2
76 Single-factor experiments—randomized block and latin square designs
where the primed totals , T'p and T' are the totals indicated without the
missing value y. In our example,
4(-2) + 4(-6)-(-12)
yij — y 33 —
(3)(3)
' '\ ^) 0 - ‘*
The reader may have wondered about a possible position effect in the prob-
lem on testing tire brands. Experience shows that rear tires get different wear
than front tires and even different sides of the same car may show different
amounts of tread wear. In the randomized block design, the four brands
were randomized onto the four wheels of each car with no regard for posi-
tion. The effect of position on wear could be balanced out by rotating the
tires every 5000 miles giving each brand 5000 miles on each wheel. However,
if this is not feasible, the positions can impose another restriction on the
randomization in such a way that each brand is not only used once on each
car but also only once in each of the four possible positions: left front, left
rear, right front, and right rear.
A design in which each treatment appears once and only once in each
row (position) and once and only once in each column (cars) is called a
Latin square design. Interest is still centered on one factor, treatments, but
two restrictions are placed on the randomization. An example of one such
4x4 Latin square is shown in Table 4.12.
Such a design is only possible when the number of levels of both restric-
tions equals the number of treatment levels. In other words, it must be a
Car
Position I II III IV
1 C D A B
2 B C D A
3 A B C D
4 D A B C
4.5 Latin squares 77
square. It is not true that all randomization is lost in this design, as the
particular Latin square to be used on a given problem may be chosen at
random from several possible Latin squares of the required size. Tables of
such squares are found in Fisher and Yates [11].
The analysis of the data in a Latin square design is a simple extension
of previous analyses where the data are now added in a third direction—
positions. If the data of Table 4.5 were imposed on the Latin square of
Table 4.12, the results could be those shown in Table 4.13.
Car
1 C 12 D 11 A 13 B 8 44
2 B 14 C 12 D 11 A 13 50
3 A 17 B 14 C 10 D 9 50
4 D 13 A 14 B 13 C 9 49
T
^ ..1
56 51 47 39 193
Treatment totals for A, B, C, and D brands are 57, 49, 43, and 44 as
before, where the model is now
and represents the positions effect. Since the only new totals are for
positions, a position sum of squares can be computed as
Source df SS MS EMS
Totals 15 80.94
that position had no significant effect, some investigators might “pool” the
position sum of squares with the error sum of squares and obtain a more
precise estimate of namely, 1.3, as given in Table 4.6. However, there
is a danger in “pooling,” as it means “accepting” a hypothesis of no position
effect, and the investigator has no idea about the possible error involved in
“accepting” a hypothesis. Naturally, if the degrees of freedom on the error
term are reduced much below that of Table 4.14, there will have to be some
pooling to get a reasonable yardstick for assessing other effects.
4.6 INTERPRETATIONS
After any ANOVA one usually wishes to investigate the single factor further
for a more detailed interpretation of the effect of the treatments. Here, after
the Latin square ANOVA of Table 4.14, we are probably concerned about
which brand to buy. Using a Newman-Keuls approach, the results are as
follows:
1. Means: 10.75 11.00 12.25 14.25
For brands : C D B A
2. = 0.9 with 6 df.
/0.9
3. Sy
J-
— = 0.47
4. For p = 2 3 4
Tabled ranges: 3.46 4.34 4.90
5. LSRs: 1.63, 2.04, 2.30.
6. Testing averages:
A-C - 3.50 > 2.30*
A-D : 3.25* > 2.04*
A-B : 2.00 > 1.63*
B-C - 1.50 < 2.04
4.7 Graeco-latin squares 79
Thus only A differs from the other brands in tread wear. A has the largest
tread wear and so is the poorest brand to buy. One could recommend
brand B, C, or D, whichever is the cheapest or whichever is “best” according
to some criterion other than tread wear.
Car
Position I II III IV
1 A(x Cy DS
2 By Ad Dec Cfi
3 C3 Dy Al^ B<x
4 DjS Ca BS Ay
Yijkm = At + ft + L + + ^ijkm
where co^ is the effect of the latest restriction with levels a, y, and S. An
outline of the analysis appears in Table 4.16.
Source df
l^i 3
L- 3
yk 3
COfti 3
^ijkm 3
Total 15
80 Single-factor experiments—randomized block and latin square designs
Such a design may not be very practical, as only 3 df are left for the
error variance.
4.8 EXTENSIONS
In some situations it is not possible to place all treatments in every block
of an experiment. This leads to a design known as an incomplete bloek
design. Sometimes a whole row or column of a Latin square is missing and
the resulting incomplete Latin square is called a Youden square. These two
special designs are discussed in Chapter 16.
4.9 SUMMARY
It might be emphasized once again that, so far, interest has been centered
on a single factor (treatments) for the discussion in Chapters 3 and 4. The
special designs simply represent restrictions on the randomization. In the
next ehapter, two or more factors are considered.
PROBLEMS
4.1 The effects of four types of graphite coaters on light box readings are to be studied.
As these readings might differ from day to day, observations are to be taken on
each of the four types every day for three days. The order of testing of the four
Problems 81
Day M A K L
Analyze these data as a randomized block design and state your conclusions.
4.2 Set up orthogonal contrasts among coater types and analyze for Problem 4.1.
4.3 Use the Newman-Keuls range test to compare four coater-type means for Problem
4.1.
4.4 If the reading on type K for the second day was missing, what missing value
should be inserted and what is the analysis now? Jud-iT 'lOioet..
4.5 Set up at least four contrasts in Problem 4.1 and use the Scheffe method to test
the significance of these.
4.6 In a research study at Purdue University on metal-removal rate, five electrode
shapes A, B, C, D, and E were studied. The removal was accomplished by an
electric discharge between the electrode and the material being cut. For this
experiment five holes were cut in five workpieces, and the order of electrodes
was arranged so that only one electrode shape was used in the same position on
each of the five workpieces. Thus, the design was a Latin square design, with
workpieces (strips) and positions on the strip as restrictions on the randomization.
Several variables were studied, one of which was the Rockwell hardness of metal
where each hole was to be cut. The results were as follows:
Position
Strip 1 2 3 4 5
Analyze these data and test for an electrode effect, position effect, and strip effect
on Rockwell hardness.
82 Single-factor experiments—randomized block and latin square designs
4.7 The times in hours necessary to cut the holes in Problem 4.6 were recorded as
follows:
} . Position
Strip 1 2 3 4 5
Analyze these data for the effect of electrodes, strips, and positions on time.
4.8 Analyze the electrode effect further in Problem 4.7 and make some statement as
to which electrodes are best if the shortest cutting time is the most desirable factor.
4.9 For an m x m Latin square, prove that a missing value may be estimated by
where the primes indicate totals for row, treatment, and column containing the
missing value and T'.. is the grand total of all actual observations.
4.10 A composite measure of screen quality was made on screens using four lacquer
concentrations, four standing times, four acryloid concentrations {A, B, C, D), and
four acetone concentrations (a, P, F, A). A Graeco-Latin square design was used
with data recorded as follows:
Lacquer Concentration
Standing Time ^ 1 li 2
4.12 In a chemical plant, five experimental treatments are to be used on a basic raw
material in an effort to increase the chemical yield. Since batches of raw material
may differ, five batches are chosen at random. Also, the order of the experiments
may affect yield as well as may the operator who performs the experiment. Con-
sidering order and operators as further restrictions on randomization, set up a
suitable design for this experiment that will be least expensive to run. Outline its
analysis.
4.13 Three groups of students are to be tested for the percentage of high-level questions
asked by each group. As questions can be on various types of material, six lessons
are taught to each group and a record is made of the percentage of high-level
questions asked by each group on all six lessons. Show a data layout for this situa-
tion and outline its ANOVA table.
N/ 4.14 Data from the results of Problem 4.13 were as follows:
Group
Lesson A B C
1 13 18 7
2 16 25 17
3 28 24 14
4 26 13 15
5 27 16 12
6 23 19 9
Time
Day 9 11 1 3 5
1 A B C D E
2 B C D E A
3 C D E A B
4 D E A B C
5 E A B C D
4.22 Write a contrast to be used in comparing the mean of fertilizers A and B with the
mean of fertilizer D and also show its standard error for Problem 4.18.
4.23 Set up three contrasts where one contrast is orthogonal to the other two but
these two are not orthogonal to each other.
4.24 a. What is the main advantage of a randomized block design over a completely
randomized design?
b. Under what conditions would the completely randomized design have been
better?
4.25 Material is analyzed for weight in grams from three vendors—A, B, C—by three
different inspectors—I, II, III—and using three different scales—1, 2, 3. The
experiment is set up on a Latin square plan with results as follows:
Problems 85
Scale
Inspector 1 2 3
I A = 16 B = 10 C = 11
II B = 15 C = 9 A = 14
III C = 13 A = n B = 13
Analyze the experiment to test for a difference in weight between vendors, between
inspectors, and between scales.
5
Factorial Experiments
5.1 INTRODUCTION
In the preceding three chapters all of the experiments involved only one
factor and its effect on a measured variable. Several different designs were
considered, but all of these represented restrictions on the randomization
where interest was still centered on the effect of a single factor.
Suppose there are now two factors of interest to the experimenter, for
example, the effect of both temperature and altitude on the current flow in a
small computer. One traditional method is to hold altitude constant and vary
the temperature and then hold temperature constant and change the altitude,
or, in general, hold all factors constant except one and take current flow
readings for several levels of this one factor, then choose another factor to
vary, holding all others constant, and so forth. To examine this type of
experimentation, consider a very simple example in which temperature is
to be set at 25°C and 55°C only and altitudes of 0 K (K = 10,000 feet) and
3 K (30,000 feet) are to be used. If one factor is to be varied at a time, the
altitude may be set for sea level or 0 K and the temperature varied from 25°C
to 55°C. Suppose the current flow changed from 210 mA to 240 mA with
this temperature increase. Now there is no way to assess whether or not this
30-mA increase is real or due to chance. Unless there is available some pre-
vious estimate of the error variability, the experiment must be repeated in
order to obtain an estimate of the error or chance variability within the
experiment. If the experiment is now repeated and the readings are 205 mA
and 230 mA as the temperature is varied from 25°C to 55°C, it seems obvious,
without any formal statistical analysis, that there is a real increase in current
flow, since for each repetition the increase is large compared with the varia-
tion in current flow within a given temperature. Graphically these results
appear as in Figure 5.1.
Four experiments have now been run to determine the effect of tem-
perature at 0 K altitude only. To check the effect of altitude, the temperature
can be held at 25°C and the altitude varied to 3 K by adjustment of pressure
86
5.1 Introduction 87
t
t yy'
Altitude
' .V/
Temperature (°C)
Altitude
Temperature (°C)
four experiments are run. Results of four such experiments might appear as
in Figure 5.3.
Figure 5.3 Factorial arrangement of temperature and altitude effect on current flow.
of a given factor are combined with all levels of every other factor in the
experiment. Thus, if four temperatures are considered at three altitudes, a
4x3 factorial experiment would be run requiring 12 different experimental
conditions. In the example above, if the current flow were 160 mA for 55°C
and 3 K, the results could be shown as in Figure 5.4.
0-A^
Altitude
0- K - #210 #240
25 55
Temperature (°C)
(a)
It is seen that the lines are much more nearly parallel and no interaction
is said to be present. An increase in temperature at 0 K produces about the
same increase in current flow (30 mA) as at 3 K (20 mA). Now a word of
warning is necessary. Lines can be made to look nearly parallel or quite
diverse depending on the scale chosen; therefore, it is necessary to run
statistical tests in order to determine whether or not the interaction is statis-
tically significant. It is only possible to test the significance of such interac-
tion if more than one observation is taken for each experimental condition.
The graphic procedure shown above is given merely to provide some insight
into what interaction is and how it might be displayed for explaining the
factorial experiment.
In addition to the advantages of a factorial experiment, it may be that
when the response variable 7 is a yield and one is concerned with which
combination of two independent variables and X2 will produce a maxi-
mum yield, the factorial will give a combination near the maximum whereas
the one-factor-at-a-time procedure will not do so. Figure 5.6 shows two
cases in which the contours indicate the yields.
^2 V2
In case (a), X2 is set and X^ is varied until the maximum yield is found,
then X^ is held there while X2 is varied and a maximum is found near the
top of the mound where the yield is around 60 units. But in case (b), the
same procedure misses the maximum by a long way, giving about a 37-unit
yield.
If, now, a factorial arrangement is used, one first notes the extremes of
each independent variable and may also choose one or more points between
these extremes. If this is done for the diagrams in Figure 5.6, Figure 5.7
shows that in both cases the maximum is reached. In some cases the maximum
may not have been reached but the factorial will indicate the direction to be
followed in the next experiment to get closer to such a maximum. This
concept is presented in Chapter 16.
5.2 Factorial experiments 91
A', ^2
Example 5.1 To determine the effect of exhaust index (in seconds) and
pump heater voltage (in volts) on the pressure inside a vacuum tube (in
microns of mercury), three exhaust indexes and two voltages are chosen at
fixed levels. It was decided to run two experiments at each of these six
treatment conditions (three exhaust indexes x two voltages). The order for
running the 12 experiments was completely randomized. This can be accom-
plished easily by labeling the six treatment conditions numbers 1 through 6
and then tossing a die to decide the order in which the experiments are to
be run.
Table 5.1 shows the results of one such complete randomization for the
12 experiments. In such a procedure some treatment combinations may be
run twice before others are run once, but this is the nature of complete
127 JJ 4 2J 1 iJ 6
9 11 12
220 4J 2 3 u 7
8 5 10
92 Factorial experiments
randomization. If one were to randomize six runs for each combination and
then randomize six more, the randomization will have been restricted and
no longer a completely randomized design. In such a procedure it is also
true that some numbers may come up again on the die even though that
combination already has been run twice. In these cases the third time is
ignored and the flipping continues until all treatment combinations are run
twice. This is, of course, a slight restriction to keep the number of observations
per treatment constant. The resulting experiment may be labeled a 3 x 2
factorial experiment with two observations per treatment with all N = 12
observations run in a completely randomized order. Of course, this random
order can be determined before the experiment is actually run and it is
important to supervise the experiment to make sure that it is actually run
in the order prescribed.
The mathematical model for this experiment can be written as
The notation k{ij) indicates that the errors are unique to each i,j combination
or are nested within each ij. Following the analysis procedures of Chapter 3,
Table 5.2 shows the resulting data in a between-treatments, within-treatments
format after multiplying all results by 1000 to eliminate decimal points.
If the data of Table 5.2 are subjected to a one-way ANOVA, the results
are as given in Table 5.3.
Fleo’ FI27 ElgO’ FI27 F/I5O, FI27 T/^Q, F22O Elgo-' F220 El 1 50’ k220
48 28 7 62 14 6
58 33 15 54 10 9
8r 106 61 22 116 24 15
lyfj. 5668 1873 274 6760 296 117
k
T. == 344
LIZ
i j k
V. == 14,988
5.2 Factorial experiments 93
Source df SS
Totals 11 5126.67
Exhaust Index
Pump Heater
Voltage 60 90 150 T-,
127 48 28 7
189
58 33 15
220 62 14 9
155
54 10 6
222 85 37 T = 344
If the sums of squares for the two main effects are computed from the
marginal totals of Table 5.4, we find
(222)2 ^ (35)2 ^ (37)2 (344)2
— = 4608.17
12
(189)2 (^55)2 (344):
SS^ = = 96.33
12
These two add to 4704.50, which is less than the sum of squares between
treatments of 4987.67. It is this difference between the treatment SS and the
SS of the two main effects that shows the interaction between the two main
effects. Here
Source df SS MS F Probability
Totals 11 5126.67
final table each main effect {El and V) can be tested for significance as well as
can the interaction by comparing each mean square with the error mean
square. Exhaust index is seen to be highly significant. Voltage is not significant
at the 5 percent significance level but the interaction is significant at the
5 percent level.
The mathematical model of Equation (5.1) can now be expanded to read
^ijk T T ^k{ij)
A
5.3 INTERPRETATIONS
Since the interaction is significant in this example, one should be very
cautious in interpreting the main effects. A significant interaction here means
that the effect of exhaust index on vacuum tube pressure at one voltage is
different from its effect at the other voltage. This can be seen graphically by
plotting the six treatment means as shown in Figure 5.8. Note that the lines
in the figure are not parallel. One reasonable procedure for interpreting such
an interaction is to run a Newman-Keuls test on the six means. Had the
main effects not interacted one could treat each set of main effect means
separately. Here testing main effects is not recommended because the results
5.3 Interpretations 95
ya.
Pressure
(X 1000)
depend on how these main effects combine. Using the six means, Newman-
Keuls gives
4. For jj.
1 p: 2 z. ^3 -r ^
If one is looking for the lowest pressure, any one of the three combinations
in the first group will minimize the pressure in the vacuum tube. Thus one
can recommend a 150-second exhaust index at either voltage or a 90-second
exhaust index at 220 volts. From a practical point of view this really gives
two choices: Pump at 127 volts for 150 seconds or at 220 volts for 90 seconds,
whichever is cheaper.
It might be noted that in the model for this problem. Equation (5.2), it
is assumed that the errors are NID (0, Ug). This means that the variances
within each of the six cells that make up the treatments are assumed to have
come from normal populations with equal variances. In the data of Table 5.4
the six ranges are 10, 5, 8, 8, 4, and 3, which average 6.33. From Table 3.15
in Chapter 3, = 20.7 and all six ranges are well within this maximum
range.
It may be of interest to note that whenever there are but two observations
per treatment, the sum of squares for the error may be computed directly.
When n = 2, the sum of squares of the ranges divided by 2 equals the SS^rror-
Here Y, = 278 and Y ^^/2 = 139, which equals the error sum of squares
of Table 5.5.
+ + + CREATE PROB3
1. 000 = 77509,113
2. 000 = COMMON (SPSS)
3. 000 = SPSS
4. 000 = #EOR
5. 000 = RUN NAME STAT 502 EX 3
6. 000 = VARIABLE LIST Y, El, PV
7. 000 = INPUT FORMAT (F2.0, IX, FEO, IX, FLO)
8. 000 = N OF CASES 12
9. 000 = VAR LABELS FI, EXHAUST INDEX/
10. 000 = PV, PUMP HEATER VOLTAGE/
11. 000 = ANOVA Y BY EI(1, 3), PV(1, 2)
12. 000 = STATISTICS ALL
13. 000 = READ INPUT DATA
14. 000 = 48 1 1
15. 000 = 58 1 1
16. 000 = 28 2 1
17. 000 - 33 2 1
18. 000 = 7 3 1
19. 000 = 15 3 1
20. 000 = 62 1 2
21. 000 = 54 1 2
22. 000 = 14 2 2
23. 000 = 10 2 2
24. 000 = 6 3 2
25. 000 = 9 3 2
26. 000 = FINISH
TOTAL POPULATION
28.67
( 12)
LI
1 2 3
55.50 21.25 9.25
( 4) ( 4) ( 4)
PV
1 2
31.50 25.83
( 6) ( 6)
LI PV
1 2
1 53.00 58.00
( 2) ( 2)
2 30.50 12.00
( 2) ( 2)
3 11.00 7.50
( 2) ( 2)
(Pp. - P...) - (Pi.. - P...) - (P.j. - P...) = l^ij. - Pi.. - P.j. + P...
If each mean is now replaced by its sample estimate, the resulting sample
model is
- y... = {Yi.. - F...) + (F,-. - Fj
+ (% - F.. - F^. + FJ + - %)
5.5 ANOVA rationale 99
If this expression is now squared and summed over i, j, and k, all cross
products vanish, and the results give
I j k i j k
i j k
+
i j k
Y,.- F,. + f,f
+ tiiiYu^-
I j k
Y^if
which again expresses the idea that the total sum of squares can be broken
down into the sum of squares between means of factor A, plus the sum of
squares between means of factor B, plus the sum of squares oi A x B
interaction, plus the error sum of squares (or within cell sum of squares).
Each sum of squares is seen to be independent of the others. Hence, if any
such sum of squares is divided by its associated degrees of freedom, the
results are independently chi-square distributed, and F tests may be run.
The degree-of-freedom breakdown would be
the interaction being cell df = {ab — 1) minus the main effect df, (a — 1)
and {b — 1) or {ab — 1) — {a — 1) — {b — 1) = ab — a — b 1 = {a — 1) x
{b — 1), and within each cell the degrees of freedom are n — \ and there
are ab such cells givjng ab{n — 1) df for error. An ANOVA table can now
be set up expanding'and simplifying the sum of squares expressions using
totals (Table 5.9).
The formulas for sum of squares in Table 5.9 provide good computa-
tional formulas for a two-way ANOVA with replication. The error sum of
squares might be rewritten as
’ n 'T'2
SS error = IZ
I J
I Yl, - ^
k n
which points up the fact that the sum of squares within each of the a x b
cells is being pooled or added for all such cells. This depends on the assump-
tion that the variance within all cells came from populations with equal
variance. The interaction sum of squares can also be rewritten as
T2
nab
100 Factorial experiments
Table 5.9 General ANOVA for Two-Factor Factorial with n Replications per Cell
Source df SS MS
a'T' 2 T' 2.
^ i.. _ ^ Each SS
Factor Ai a - 1
^ nb nab divided by
b'J'2 rp 2 its df
^ -j- ^ ...
Factor Bj b - 1 ^ na nab
d b -y-t 2 ^ 2
^ O'- _ ^ ^■
A X B interaction {a - !)(/) - 1)
ij n 1 nb^
b rp2 7^ 2
y ^ T...
j na nab
abn a b rj-,2
Error ab{n — 1)
i j k i j ^
abn rp2
Totals abn — 1
? ? ? -...
Example 5.2 To extend the factorial idea a bit further, consider a problem
with three factors. Such a problem was presented in Chapter 1 on the effect
of tool type, angle of bevel, and type of cut on power consumption for ceramic
tool cutting. Reference to this problem will point out the phases of experi-
ment, design, and analysis as followed in Example 5.1.
It is a 2 X 2 X 2 factorial experiment with four observations per cell
run in a completely randomized manner. The mathematical model is
Table 5.10 Coded Ceramic Tool Data of Table 1.2, Code: 2(X-28.0)
Tool Type
1 2
Continuous 2 10 3
5 (4^ 4 0 2 25
-2 9 -6 0
2 15 -5 13
Interrupted 0 —2 —7 —1
-6 2 -6 0
-3 (^54j -1 nm 0 -2 (21) -38
-3 -1 -4 -4
-12 -2 -17 -7
Tool type sum of squares: Add for each tool type. The totals are +3
and —16; square these and divide by the number of observations per type
(16), add these results for both types, and subtract the correction term. Thus,
3^ + (-16)^ (-13)^
SS tool type 11.28
16
Bevel angle sum of squares: Same procedure on the totals for each bevel
angle, — 32 and 19:
Type of cut sum of squares: Same procedure, with cut totals of 25 and — 38
For the T x B interaction, ignore type of cut and use cell totals for the
T X B cells. These are —10, 13, —22, +6:
- 0.78
For T X C interaction, ignore bevel angle and the cell totals become 17,
-14, 8, -24:
(17)2+ (-14)2+ (8)2+ (-24)2 (-13)^
SS T X C interaction - 11.28 - 124.03
8 32
= 0.03
For B X C interaction, ignore tool type and the cell totals become —3, 28,
-29, -9:
(- 3)2 + (28)2 (_ 29)2 + (_ 9)2 (_ 13):
SS B X C interaction -81.28- 124.03
8 32
= 3.78
X B X C interaction
(-13)2
- ^ - 11.28 - 81.28 - 124.03 - 0.78 - 0.03 - 3.78 = 0.79
By subtraction
SS error 213.75
These results are displayed in Table 5.11. If they are compared with those
in Table 1.3, they appear to differ considerably. Actually they give the same
F test results, but in Table 5.11 the data were coded involving multiplication
by 2; the data of Table 1.3 are uncoded. Multiplication by 2 will multiply
the variance or mean square by 4; thus if all mean squares in Table 5.11 are
divided by 4, the results are the same as in Table 1.3. For example, on tool
types 11.28/4 = 2.82, and error 213.75/4 = 53.44. It is worth noting that
any decoding is unnecessary for determining the F ratios. However, if one
wishes confidence limits on the original data or components of variance on
the original data, it may be necessary to decode the results.
5.6 Remarks 103
Source df SS MS
Totals 31 435.72
5.6 REMARKS
Since the examples in this chapter have contained several replications within
a cell, it would be well to examine a situation involving only one observation
per cell. In this case k = 1, and the model is written as
Yij = fd Ai + Bj + ABij + 8ij
A glance at the last two terms indicates that we cannot distinguish between
the interaction and the error—they are hopelessly confounded. Then the
only reasonable situation for running one observation per cell is one in
which past experience generally assures us that there is no interaction. In
such a case, the model is written as
Yij = jd + Ai Bj Y Sij
It may also be noted that this model looks very much like the model for a
randomized block design for a single-factor experiment (Chapter 4). In
Chapter 4 that model was written as
Yij = + Pi Y Sij
Even though the models do look alike and an analysis would be run in
the same way, this latter is a single-factor experiment—treatments are the
factor—and pi represents a restriction on the randomization. In the factorial
104 Factorial experiments
model, there are two factors of interest, Ai and Bj, and the design is com-
pletely randomized. It is, however, assumed in the randomized-block situa-
tion that there is no interaction between treatments and blocks. This is often
a more reasonable assumption for blocks and treatments since blocks are
often chosen at random. For a two-factor experiment an interaction between
A and B may very well be present, and some external information must be
available in order to assume that no such interaction exists. An experi-
menter who is not sure about interaction must take more than one observa-
tion per cell and test the hypotheses of no interaction.
As the number of factors increases, however, the presence of higher order
interactions is much more unlikely, so it is fairly safe to assume no four-way,
five-way,..., interactions. Even if these were present, they would be difficult
to explain in practical terms.
5.7 SUMMARY
I. Single factor
1. Completely randomized 1. One-way ANOVA
Y,j = n + Tj + Cij
2. Randomized block 2.
Tj = + Tj + Pi + &ij
a. Complete a. Two-way ANOVA
b. Incomplete, balanced b. Special ANOVA
c. Incomplete, general c. Regression method
3. Latin square
Tjfc = g + + Tj
+ 7fc + fitjTc
a. Complete a. Three-way ANOVA
b. Incomplete, b. Special ANOVA
Youden square (like 2b)
4. Graeco-Latin square 4. Four-way ANOVA
Tjfcm = M + ^( + L
+ Tic + + ^ijkm
11. Two or more
factors
A. Factorial
(crossed)
1. Completely randomized 1.
Tjfc — ^ + Ai + Bj
+ ABij + •••
for more factors
a. General case a. ANOVA with
interactions
Problems 105
PROBLEMS
5.1 To determine the effect of two glass types and three phosphor types on the light
output of a television tube, light output is measured by the current required in
series with the tube to produce 30 foot-lamberts of light output. Thus the higher
the current is in microamperes, the poorer the tube is in light output. Three ob-
servations were taken under each of the six treatment conditions and the experi-
ment was completely randomized. The following data were recorded.
Phosphor Type
Glass Type A B C
Do an analysis of variance on these data and test the effect of glass type, phosphor
types, and interaction on the current flow.
5.2 Plot the results of Problem 5.1 to show that your conclusions are reasonable.
5.3 For any significant effects in Problem 5.1 test further between the levels of the
significant factors.
5.4 Based on the results in Problems 5.1-5.3, what glass type and phosphor type
(or types) would you recommend if a low current is most desirable?
5.5 Adhesive force on gummed material was determined under three fixed humidity
and three fixed temperature conditions. Four readings were made under each set
of conditions. The experiment was completely randomized and the results set out
in an ANOVA table as follows:
Source df SS MS
Humidity 9.07
Temperature 8.66
H X T interaction 6.07
Error
Total 52.30
and two materials with two samples tested under each set of conditions. The order
of the experiment is completely randomized and the levels of all factors are fixed.
The following data are recorded on thrust forces after subtracting 200 from all
readings.
Speed
Temperature (°C)
145 155 165
5.12 What is there in the results of Problem 5.11 that would lead you to question
some assumption about the experiment?
5.13 Tomato plants were grown in a greenhouse under treatments consisting of com-
binations of soil type (factor A) and fertilizer type (factor B). A completely ran-
domized two-factor design was used with two replications per cell. The following
data on the yield Y (in kilograms) of tomatoes were obtained for the 30 plants
under study.
Soil
Type{A) 1 2 3 T,
5.14 Carry out tests on main effects and interactions in an appropriate order using
levels of significance for each test that seem appropriate to you. For each test
state HQ and show the rejection region, and give your conclusion. After
doing all tests, summarize your results in words.
5.15 On the basis of your results in Problem 5.14, state how you would carry out
Newman-Keuls comparisons to find out the best combination (or combinations)
of soil type and fertilizer type to achieve maximum mean yield of tomatoes. Then
carry out these comparisons (at an a of 0.05) and state your conclusions.
5.16 Each of the following tables represents the cell means (7^^ ) for a two-factor com-
pletely randomized balanced experiment. For each table tell which (if any) of the
following SS’s would be zero for that table: SS^, SS^, SS^g. (There may be none;
there may be more than one.)
a. A
1 3
5 3
108 Factorial experiments
b. A
= 0
c.
= 0
d.
= 0
5.17 In each two-factor design table below the numbers in the cells of the table are the
population means of the observations for those cells. For each table, indicate
whether or not there is interaction between the factors and justify your assertion.
a. Factor B
1 2 3 4
Factor A 1 7 6 5 2
2 9 8 7 4
b. Factor B
1 2 3
Factor A 1 7 6 5
2 5 6 7
c. Factor B
5.18 An industrial engineer presented two types of stimuli (two-dimensional and three-
dimensional films) of two diflferent jobs (1 and 2) to each of hve analysts. Each
analyst was presented each job-stimulus film twice and the order of the whole
experiment was considered completely randomized. The engineer was interested
in the consistency of analyst ratings of four sequences within each job stimulus
presentation. The variable recorded is the log variance of the four sequences
since log variance is more likely to be normally distributed than the variance.
Data showed:
Stimulus
T wo-Dimensional Three-Dimensional
Source df SS
Sex 1 239,763
Size of school 2 423,056
Years in position 5 564,689
Sex by size interaction 2 18,459
Sex X years interaction 5 85,901
Size X years interaction 10 240,115
Sex X size x years interaction 10 151,394
Within classes 153 1,501,642
Indicate the mathematical model for the above study. Show a possible data layout.
Complete the ANOVA table and comment on any significant results.
5.21 Four factors are studied for their effect on the luster of plastic film. These factors
are (1) film thickness (1 or 2 mils); (2) drying conditions (regular or special); (3)
110 Factorial experiments
length of wash (20, 30, 40, or 60 minutes); and (4) temperature of wash (92°C or
100°C). Two observations of film luster are taken under each set of conditions.
Assuming complete randomization, analyze the following data.
1-mil Thickness
20 3.4 3.4 19.6 14.5 2.1 3.8 17.2 13.4
30 4.1 4.1 17.5 17.0 4.0 4.6 13.5 14.3
40 4.9 4.2 17.6 15.2 5.1 3.3 16.0 17.8
60 5.0 4.9 20.9 17.1 8.1 4.3 17.5 13.9
2-mil Thickness
20 5.5 3.7 26.6 29.5 4.5 4.5 25.6 22.5
30 5.7 6.1 31.6 30.2 5.9 5.9 29.2 29.8
40 5.6 5.6 30.5 30.2 5.5 5.8 32.6 27.4
60 7.2 6.0 31.4 29.6 8.0 9.9 33.5 29.5
5.22 Plot and discuss any significant interactions found in Problem 5.21. What condi-
tions would you recommend for maximum luster and why?
5.23 To study the effect of four types of wax applied to floors for three different lengths
of polishing times it was decided to try each combination twice and completely to
randomize the order in which ffoors received which treatments. The criterion was
a gloss index.
a. Outline an ANOVA table for this problem with proper degrees of freedom.
b. Explain, in words, what a significant interaction would mean in this situation.
5.24 A manufacturer of combustion engines is concerned about the percentage of smoke
emitted by the engine in order to meet EPA standards. An experiment was con-
ducted involving engines with three timing levels (T), three throat diameters (D),
two volume ratios in the combustion chamber (K), and two injection systems (/).
Thirty-six engines were designed to represent all possible combinations of these
four factors and the engines were then operated in a completely random order
and percent smoke recorded to the nearest 0.1 percent. Results showed:
1 2 3
D D D
1 2 3 1 2 3 1 2 3
5.25 Students’ scores on a psychomotor test were recorded for various sized targets at
which the student must aim, different machines used, and the level of background
illumination. Results are shown in the following ANOVA table.
Source df SS MS
6.1 INTRODUCTION
In the last chapter factorial experiments were considered and a general
method for their analysis was given. There are a few special cases that are
of considerable interest in future designs. One of these is the case of / factors
where each factor is at just two levels. These levels might be two extremes of
temperature, two extremes of pressure, two time values, two machines, and
so on. Although this may seem like a rather trivial case since only two
levels are involved, it is, nevertheless, very useful for at least two reasons:
to introduce notation and concepts useful when more involved designs are
discussed and to illustrate what main effects and interactions there really
are in this simple case. It is also true that in practice many experiments are
run at just two levels of each factor. The ceramic tool-cutting example of
Chapters 1 and 5 is a 2 x 2 x 2, or 2^ factorial, with four observations per
cell. Throughout this chapter the two levels will be considered as fixed levels.
This is quite reasonable because the two levels are chosen at points near the
extremes, rather than at random.
6.2 2^ FACTORIAL
The simplest case to consider is one in which two factors are of interest and
each factor is set at just two levels. This is a 2 x 2 = 2^ factorial, and the
design will be considered as completely randomized. The example in Sec-
tion 5.1 is of this type. The factors are temperature and altitude, and each
is set at two levels: temperature at 25°C and 55°C and altitude at 0 K and
3 K. This gives four treatment combinations, displayed in Figure 5.3 where
the response variable is the current flow in milliamperes. To generalize a bit,
consider temperature as factor A and altitude as factor B. The model for
this completely randomized design would be
112
6.2 2^ Factorial 113
/? = 180 ab = 200
Factor B
(altitude)
114 2^ Factorial experiments
high level of each factor with the low level of the other {a, b), and the fourth
term is the algebraic product of the second and third (ab). When a third
factor is introduced, it is placed at the end of this sequence and then mul-
tiplied by all of its predecessors. For example, if factor C is also present at
two levels, the treatment combinations are
Note that the coefficients on this A effect are all -I-1 when A is at its high
level in the treatment combinations ( -\-a, -\-ab) and the coefficients are all
— 1 when A is at its low level as in (1) and b. Note also that
2A = —(l)-\-a — b + ab
B = j\^b — (1) + ab — a^
or
B — — — a + b + uh]
and again it is seen that the same four responses are used, but + 1 coefficients
are on the treatment combinations for the high level of B and — 1 coefficients
for the low level of B. Also, 2B is a contrast.
To determine the effect of the interaction between A and B, note that at
the high level of B the A effect is ab — b, and at the low level of B the A
effect is a — (1). If these two effects differ, there is an interaction between
+ and B. Thus the interaction is the average difference between these two
differences
AB = i{(uh - b) - [a - (1)]}
= j\^ab — b — a -\- (1)]
or
= i[(l) “ ^
6.2 2^ Factorial 115
Here again the same four treatment combinations are used with a
different combination of coefficients. Note that 2AB is also a contrast, and
all contrasts 2A, 2B, 2AB are orthogonal to each other. Summarizing in
normal order gives
2A = —(1) -\- a — b + ab
2B = —{1) — a b ab
2AB = — a — b + ab
Note that the interaction effect takes the responses on one diagonal of the
square with +1 coefficients and the responses on the other diagonal with
— 1 coefficients. Note also that the coefficients for the interaction effect can
be found by multiplying the corresponding coefficients of the two main
effects. As the only coefficients used are + I’s and — I’s, the proper coefficients
on the treatment combinations for each main effect and interaction can be
determined from Table 6.1.
Treatment Effect
Combination A B AB
(1) — — +
a + — —
b — + —
ab + + +
From Table 6.1 the orthogonality of the effects is easily seen, as well as
the generation of the interaction coefficients from the main effect coefficients.
Another approach to the interaction between A and B would be to
consider that at the high level of A the effect of B is ab — u, and at the low
level of A the effect of 5 is 6 — (1), so that the average difference is
AB = j{{ab — u) — [6 — (1)]}
= ^\_ab — a — b + (1)]
^ i[ + (l) ~ ^ ~ ^
Since 2A, 2B, and 2AB are contrasts, the sum of squares due to a contrast is
(contrast)^
SS
SS, . . as
4
r2(-35)P
SSfi = = 1225
4
SS,. . . 25
= 1875
Since each effect and the interaction has but 1 df and there is no measure of
error as only one observation was taken in each cell, this ANOVA is quite
trivial, but it does show another approach to analysis based on effects. The
simple ANOVA table would be that shown by Table 6.2.
Source df SS MS
At 1 625 625
fi; 1 1225 1225
Error or AB^j 1 25 25
Totals 3 1875
If the general methods of Chapter 5 are used on the data in Figure 6.1
(coded by subtracting 200), the results are those in Table 6.3.
(30)^
SS.otal = (10)" + (-20)2 + (40)2 _ = 1875
Factor A
Factor B 0 1 Totals
0 + 10 + 40 + 50
1 -20 0 -20
Totals -10 + 40 + 30
which are the same sums of squares as given in Table 6.2. Even though no
separate measure of error is available, a glance at the mean squares of
Table 6.2 shows that both main effects are large compared with the inter-
action effect.
The results of this section can easily be extended to cases where there
are n replications in the cells, by using cell totals for the responses at (1), a,
b, and ab and adjusting the sum of squares for the effects accordingly (see
Example 6.1).
(10)"
= 170 - 12.5 = 157.5
8
(20)" + (-10)^
SS 12.5 = 112.5
(10)" + (0)'
SS^ = 12.5 = 12.5
= 12.5
(1) = 10 a = 0 b = 10 ab = —10
118 2^ Factorial experiments
Factor A
Factor B 0 1 Totals
0 4 2
6 -2
10 0 + 10
1 3 -4
7 -6
10 -10 0
Totals + 20 -10 + 10
The coefficient 4 used with each response represents the two individual
responses at each level. In general the coefficient of these effects is n • 2^“^
where n is the number of replications and / the number of factors. Here
/ = 2 and n = 2. The sum of squares for these three contrasts are
1 + 1 + 1 + 1 = 4
i= 1
These results are the same as given by the general method. The total
sum of squares must be calculated as usual in order to get the error sum of
squares by subtraction.
For this special case of a 2^ factorial experiment, Yates [26] developed
a rather simple scheme for computing these contrasts. The method can best
be illustrated on Example 6.1, using Table 6.5.
6.2 1? Factorial 119
Treatment
Combination Response (1) (2) SS
In Table 6.5 list all treatment combinations in the first column. Place
the total response to each of these treatment combinations in the second
column. For the third column, labeled (1), add the responses in pairs, for
example, 10 + 0 = 10 for the first two and 10 — 10 = 0 for the next two;
this completes half of column (1). For the second half, subtract the responses
in pairs, always subtracting the first from the second, for example, 0 — 10 =
— 10; —10 — (10) = —20. This completes column (1). Column (2) is deter-
mined in the same manner as column (1) using the column (1) results:
10 + 0-10; -10 - 20 - -30; 0 - 10 - -10; -20 - (-10) - -10.
Proceed in this same manner until the/th column is reached: (1), (2), (3),...,
(/). In this case / — 2, so there are just two columns. The values in column
/ are the contrasts, where the first entry is the grand total of all readings; the
one corresponding to u is n • 2^~ ^ ‘ A,b is n ’ 2^~ ^ • B, and ab is n • 2^~ ^ • AB.
When the results of the last column [column (/)] are squared and
divided by n • 2^, the results are the sums of squares as shown above. The
first sum of squares (total)^/8 is the correction term for the grand mean
given in Chapter 5. The Yates method reduces the analysis to simply adding
and subtracting numbers. It is very useful provided the experiment is a 2^
factorial with n observations per treatment.
As proof for the Yates method on a 2^ factorial go through the steps
above using the treatment combination symbols for the responses as in
Table 6.6. It is obvious that the last column does give the proper treatment
contrasts.
Treatment
Combination (1) (2)
6.3 V FACTORIAL
Considering a third factor C, also at two levels, the experiment will be a
2 X 2 X 2 or 2^ factorial, again run in a completely randomized manner.
The treatment combinations are now (1), u, b, ab, c, ac, be, abc, and they may
be represented as vertices of a cube as in Figure 6.2.
Factor B
Factor A
4A = ~{1) a — b ab — c ac — be abc
For factor B consider responses in the lower and higher planes of the cube
4B — —{!) — a b A ab — c — ac + be + abc
4C = c A be A ac A abc — {1) — b — a — ab
or
4C = —{1) — a — b — ab A c A ac A be A abc
The AC and BC interactions can be determined as AB was, and it can be
seen that the resulting interaction effects are
Their difference is
which can also be obtained from multiplying the coefficients of A and BC,
or B and AC, or C and AB. These others can also be used to get the ABC
interaction, but the results are the same. Summarizing gives us Table 6.7.
Table 6.7 illustrates once again the orthogonality of the effects and can
easily be extended to four, five, and more factors if each factor is at two
levels only. In a 2^-factorial experiment, the sum of squares is given by
(contrast)^ (contrast)^
^^contrast ^3 o
n • 2 8n
and again the Yates method leads to an easy computation of the contrasts.
122 2^ Factorial experiments
Effect
Treatment
Combination Total A B AB C AC BC ABC
(1) + — — + — + + —
a + + — — — — + +
b + — + — —
+ — +
ab + + + + — — — —
c + — —
+ + — — +
ac + + — — + + — —
be + — + — + — + —
abc + + + + + + + +
Tool Type
1 >
Continuous (1) b a ab
2 15 -5 13
Interrupted c be ae abe
-12 -2 -17 -7
In Table 6.8 the totals of the four observations for each treatment
combination have been entered in their corresponding cells. By the Yates
method we get the entries in Table 6.9.
The sum of squares is seen to be in substantial agreement with those
of Table 5.10. We must resort to the individual readings, however, to deter-
mine the total sum of squares and then the error sum of squares. This was
done in Table 5.10 and the interpretation of this problem is given in Chapter 5.
The purpose of repeating the problem here was merely to show the use of
the Yates method on a 2^-factorial experiment.
6.4 2^—Remarks 123
Treatment
Combination Response (1) (2) (3) SS
6.4 2^—REMARKS
Source df
Main effects A 1
B 1
if
C 1
• * ^
Two-factor interactions AB 1
AC 1 /(/-I)
C(/, 2) = i
BC 1
♦ • >
Total n-2^ - i
124 2^ Factorial experiments
effect = — (contrast)
nz-'
and
(contrast)^
SS contrast
n2-^
6.5 SUMMARY
I. Single factor
1. Completely randomized 1. One-way ANOVA
Yij = + Tj + Sij
2. Randomized block 2.
Tj = /^ + +C+ ^ij
a. Complete a. Two-way ANOVA
b. Incomplete, balanced b. Special ANOVA
c. Incomplete, general c. Regression method
3. Latin square
Yijk = /^ + A + C
+ Tfc + ^ijk
a. Complete a. Three-way ANOVA
b. Incomplete, b. Special ANOVA
Youden square (like 2b)
4. Graeco-Latin square 4. Four-way ANOVA
Yijkm = M + C
+ + ^ijkm
II. Two or more
factors
A. Factorial
(crossed)
1. Completely randomized
^ijk = + Ai Bj
+ ABij + e;t(U) ■ ■ ■
for more factors
a. General case a. ANOVA with
interactions
b. 2^ case b. Yates’ method or
general ANOVA;
use (1), a, b, ah, . . .
Problems 125
PROBLEMS
6.1 For the 2^ factorial with three observations per cell given below (hypothetical
data), do an analysis by the method of Chapter 5.
Factor A
Factor B ^1 ^2
0 4
2 6
1 2
B2 -1 -1
-3 -3
1 -7
6.2 Redo Problem 6.1 by the Yates method and compare results.
6.3 In an experiment on chemical yield three factors were studied, each at two levels.
The experiment was completely randomized and the factors were known only
as A, B, and C. The results were
A1 A2
B, B2 B, B'l
c, C2 Cl C2 Cl C2 c, C2
1595 1745 1835 1838 1573 2184 1700 1717
1578 1689 1823 1614 1592 1538 1815 1806
5! ''3 3434 365^ 3463 3^^ 2 3 '■ 6 '9-.; ,
Analyze by the methods of Chapter 5. •n' . .
6.4 Analyze Problem 6.3 by the Yates method and compare your results.
6.5 Plot any results from Problem 6.3 that might be meaningful from a management
point of view.
6.6 The results of Problem 6.3 lead to another experiment with four factors each at
two levels, with the following data.
^1
Bi B2
Cl C2 Cl C2
^1 D2 D, Di D, D2 D2
1985 2156 1694 2184 1765 1923 1806 1957
1592 2032 1712 1921 1700 2007 1758 1717
'i -( ■ - -
126 2^ Factorial experiments
Ai
B2
Cl C2 Cl C2
D, D2 D, D2 D, D2 D2
. € 3- 2%3^
Analyze these data by the general methods of Chapter 5.
6.7 Analyze Problem 6.6 using the Yates method.
6.8 Plot from Problem 6.6 any results you think are meaningful.
6.9 Consider a four-factor experiment with each factor at two levels.
a. Write out the treatment combinations in this experiment in normal order.
b. Assuming there is only one observation per treatment, outline an ANOVA for
this example and argue what you would use for an error term to test for signifi-
cance of various factors and interactions.
c. Explain how you would determine the AD interaction from the responses to
the treatment combinations given in part (a).
6.10 An experimenter is interested in the effects of five factors:—A, B, C, D, and E—
each at two levels on some response variable Y. Assuming that data are available
for each of the possible treatment combinations in this experiment, answer the
following.
a. How would you determine the effect of factor C and its sum of squares based
on data taken at each of the treatment combinations?
b. Set up an ANOVA table for this problem assuming n observations for each
treatment combination.
c. If n = 2 for part (b), at what numerical value of F would you reject the hypoth-
esis of no ACE interaction.
d. How would you propose testing for ACE interaction and what F would be
necessary for rejection in this problem if n = 1 ?
6.11 Consider an experiment with five factors A, B, C, D, and £, each at two levels.
a. How many treatment combinations are there?
b. Outline an ANOVA table to test all main effects, all two-way and all three-way
interactions, if there is only one observation per treatment combination.
c. Explain how you would find the proper signs on the treatment combinations
for the ACE interaction.
d. Some experimenter comments to you that “since all factors are at only two
levels, it is unnecessary to run a Newman-Keuls test as one only has to look
and see which level gives the greater mean response.” Explain situations in
which this comment would be true for your results and in which situations it
would not be true.
6.12 Three factors are studied for their effect on the horsepower necessary to remove a
cubic inch of metal per minute. The factors are feed rate at two levels, tool condition
Problems 127
at two levels, and tool type at two levels. For each treatment combination three
observations are taken and all 24 observations of horsepower are made in a
completely randomized order. The data appear below.
Feed Rate
0.011 0.015
6.13 If minimum horsepower is desired for the experiment given in Problem 6.12,
what are your recommendations as to the best set of factor conditions on the basis
of your results in Problem 6.12.
6.14 Any factor whose levels are powers of 2 such as 4, 8,16,..., may be considered as
pseudo factors in the form 2^, 2^, 2"^,.... Yates’ method can then be used if all
factor levels are 2 or powers of 2 and the resulting sums of squares for the pseudo
factors can be added to give the sums of squares for the actual factors and appro-
priate interactions. Try out this scheme on the data of Problem 5.21.
6.15 A systematic test was made to determine the effects on the coil breakdown voltage
of the following six variables, each at two levels as indicated.
1. Firing furnace: Number 1 or 3
2. Firing temperature: 1650°C or 1700°C
3. Gas humidification: Yes or no
4. Coil outside diameter: Large or small
5. Artificial chipping: Yes or no
6. Sleeve: Number 1 or 2
Assuming complete randomization, devise a data sheet for this experiment, outline
its ANOVA, and explain your error term if only one coil is to be used for each of
the treatment combinations.
6.16 If seven factors are to be studied at two levels each, outline the ANOVA for this
experiment in a table similar to Table 6.10, and, assuming n = 1, suggest a reason-
able error term.
128 2^ Factorial experiments
6.17 In a steel mill an experiment was designed to study the effect of drop-out tempera-
ture, back-zone temperature, and type of atmosphere on heat slivers in samples of
steel. The response variable is a mean surface factor obtained for each type of
sliver. Results are
Drop-out Temperature
2145° 2165°
Back-Zone Temperature
Drop-out Temperature
2145° 2165°
Back-Zone Temperature
For these data the SS^rror = 13.22 based on three repeated measurements. Com-
pletely analyze these data.
6.20 Sketch rough graphs for the following situations where each factor is at two
levels only:
1. Main effects A and B both significant, AB interaction not significant.
2. Main effect A significant, B not significant, and AB significant.
3. Both main effects not significant but interaction significant.
6.21 For a 2^ factorial can you sketch a graph (or graphs) to show no significant two-
factor interactions, but a significant three-factor interaction?
7
Qualitative and Quantitative Factors
7.1 INTRODUCTION
In Chapter 3, Figure 3.1, we noted several possible procedures to follow
after an analysis of variance. In that chapter fixed levels of a single factor
were considered when such levels were qualitative levels. If the levels are
quantitative, such as temperature, dosage, humidity, or time, a different
route is suggested after an ANOVA yields significant differences between
response variable averages. It is also noted that the path to follow depends
upon whether or not these quantitative levels are equispaced. It will be seen
in this chapter that there are definite advantages in designing the experiment
so that the levels of any quantitative variables are equispaced.
It is assumed that the reader has had some introduction to linear re-
gression in which one tries to write a mathematical model of a straight line
that will predict some response variable Y from one independent variable
X. After reviewing this linear “fit,” the concept will be extended to fitting
a curve when a straight line is not an adequate model. These notions will
then be extended using the method of orthogonal polynomials to test for the
highest order polynomial necessary to do a decent prediction of Y from X.
The analysis will be made only if the analysis of variance shows a significant
difference in the Y averages due to levels of X and we wish to go further
and see how Y may be related to X.
After building some methodology to pull out various trend lines, the
notions will be extended to more than one factor where the factors in the
experiment may be all at quantitative levels or where some factors may be
at qualitative levels and some at quantitative levels. The next chapter will
extend these ideas to a multiple regression model where prediction equations
are sought and the experiment may not necessarily have all quantitative
levels equispaced.
129
130 Qualitative and quantitative factors
To review linear regression and to become familiar with the notation used,
consider an example where an experiment was designed to determine the
effect of the amount of drug dosage on a person’s reaction time to a given
stimulus.
Dosage (mg)
0.5 1.0 1.5 2.0 2.5
Reaction 26 28 28 32 38
Time (ms) 28 26 30 33 39
yu 29 30 31 31 38
n 3 3 3 3 3 N = 15
83 84 89 96 115 T . = 467
Tj
lyfj 2301 2360 2645 3074 4409 XX yfj = 14,789
i i j
Source df SS MS F Significance
Totals 14 249.73
a line is shown in Figure 7.1. One can see how close the line comes to the
average Y values for each X. Note that for a given Xj (say 2.5 mg) the observed
Yij can be partitioned into three parts: Y'x is the predicted value of Y^j for a
given Xj as found on the straight line, Yj — Y'x is the amount by which the
mean Y for a given Xj departs from its predicted value Y'x (referred to as
the departure from linear regression in this model), and Y^j — fj is the
amount by which an observed Y^j varies from its treatment mean.
In the analysis of variance model
and now one tries to predict fij from some function based on Xj. If this
predicted mean is labeled /ty/x, the model can be expanded to read
where jiy/x is the true predicted mean based on X and fij — fiy/x is the amount
by which the true mean juj departs from its predicted value.
For the sample model Equation (7.2) becomes
If the second and third terms on the right of this equation are combined, we
have the ANOVA sample model. Thus, we attempt here to partition the
132 Qualitative and quantitative factors
mean of a given treatment into two parts: that which can be predicted by
regression on X and the amount by which the mean shows a departure from
such a regression model The purpose, of course, is to find a model for pre-
dicting the response means such that the departures from these means are
very small By choosing a polynomial of high enough degree one can find
a model that actually goes through all the Y means. However, it is hoped
that an adequate lower degree model can be found whose departures are
small and insignificant when compared with the error of individual T’s
around their means.
In Equation (7.3) Y'x represents the predicted Y for a given X for any
assumed model. The straight-line model is only one possible attempt to “fit”
a curve to the data. The straight-line model is given by
= feo + (7.4)
where is the Y intercept and is its slope. The usual method for deter-
mining the values of BQ and is the method of least squares. The values of
bg and b^ are determined from the data in such a way as to minimize the
sum of squares of the deviations of each from its predicted value Y'x.
[One could also find bg and b^ such that the sum of squares of the departures
from regression {Yj — T^) is minimized, but this leads to the same results.]
Here
Yij - n = Y,j - foo - b,Xj
and
= y y {Y,j - yy^ = yX (Y^J -to- b,x/
i j i j
Since the summations of X’s and T’s can be found from a given set of
data, SSdeviations is a function of bg and b^.To minimize SS^eviations then, one
differentiates SS^eviations with respect to bg and b^
^ 2y y (1^. - bo - b,Xj){-Xj) = 0
OOi i j
Y.I.Y,J = boN +
i j j
These equations are often called the least squares normal equations, which
^an now be solved for bg and
7.2 Linear regression 133
^ 467 - (22.5)(5.07)
bo = = 23.53
or the ratio of the sum of cross products of X and Y to the sum of the squares
of X. And 6o as
bo = Y - b^X
Here
I- Y VVj
Zi I; 2^1 li YJ
SP XY N
hi =
SS X Hi 1J XJ}'-
Y IJ Xj - N
134 Qualitative and quantitative factors
as before and
467
= Y - = (5.07) = 23.53
U
as before.
Totals 0 12.401
To see how well this linear model predicts the T’s from the X’s, we
construct Table 7.3. With the predicted values taken from the model
= 23.53 + 5S)lXj
The table shows the departures from linear regression, for example,
7j- - 7^ = 28.000 - 28.600 = -0.600
Note that the sum of the departures adds to zero and the sum of squares
of departures from linear regression is
Source df SS MS F Significance
Totals 14 249.73
Thus linear regression will account for about 77 percent of the variation
seen in the reaction time Y.
By taking the ratio of the sum of squares between means to the total
sum of squares, one finds the maximum amount of the total variation that
could be accounted for by a curve or model that passes through all the
mean T’s for each X. This is called eta squared {q^). Here
= 229.73/249.73 = 0.9199
This expression adds the error about the mean and the departure of the
mean from the predicted curve, which is appropriate only if the departure is
136 Qualitative and quantitative factors
nonsignificant. Then
^^departure T SS error
>y.x N - 2
'37.20 + 20.00
^Y.x — = 2.10
13
This statistic is often used to set confidence limits around a line of best fit
when linear regression is appropriate.
In this discussion of Example 7.1 no use has been made of the fact that
the dosages {XjS) are equispaced. When this is true, one can code the XjS
by considering their mean X and the width of the interval between them
c. Let
Xj-X, Xj - 1-5
Uj =
0.5
and our model becomes
n = ^0 + b\u (7.6)
b'0 I. Z. Yi = Y
N
(7.8)
u, ^ ^ Zj "jTj
2
n ZJ “j ” Z; “ j J
and they are relatively easy to solve. Note, too, that since Yj
expression Yj ^j^.j ^ contrast and its sum of squares is (Yj UjT j)^/n Y
For our data
^0 Y = 467/15 = 31.133
as before. Note also that the sum of squares due to linear regression can be
found from the linear contrast
(76)^
SS due to linear - 192.53
regression 3(10)
as shown in Table 7.4.
These equations can be solved for ho, b^, and 62, but it is difficult. Again
coded by w/s,
Y'„ = b'o + b\uj + b'^uj (7.11)
Z Z Yij =
I J
b’oN + b'lii Z Uj + fc'jn Z
Z Z njYij
i j
= b'on Z Uj +
j
b\n'^ u] + b'^ri
j
Z u]
j
(7.12)
Z Z njYij
1 J
= b'oii Z uj + b\n Z uj + b'^n Z uj
Because of the choice of the u/s, the sums of all odd powers of the w/s are
zero (^ Uj = ~ equations become
ZZ
I J
= b'oN + b'^n Z uj
Z Z njYij
i j
= b\nY, uj
J
(7.13)
For the data of Example 7.1, the middle equation of (7.13) gives
b\ = E "ija 2.53
n Uj 30
The other two equations of (7.13) give:
and
X uf = (-2)^ + (-1)^ + (0)^ + (1)^ + (2)^ - 34
j
and
y.j
Y' Yj - Y', {y.j - yx
' )"
One need not write this in terms of Xj to see how well it predicts the y’s
from the X’s. Table 7.5 shows the predicted T’s and the departure of the means
from the quadratic model.
From Table 7.5 the sum of the squares of the departure from the qua-
dratic is
and this step can be added to refine Table 7.4 further and give Table 7.6.
Now the departure from the quadratic is not significant at the 5 percent level
so it is appropriate to stop with the quadratic for predicting reaction time
from dosage. One may also note from Table 7.6 that linear and quadratic
terms together account for 192.53 + 34.38 = 226.91 of the total sum of
squares in Y of 249.73. This is 226.91/249.73 = 0.9086 or approximately
91 percent of the variability accounted for. The square root of this statistic
is sometimes referred to as a correlation ratio R and = 0.91. The error
and nonsignificant departure term could be ‘pooled’ to give a standard error
of estimate for the quadratic
'20.00 + 2.82
^Y.X, 1.38
10 + 2
which might be used to set confidence limits on the T’s around the quadratic
curve.
Source df SS MS F Significance
Totals 14 249.73
+ ^1^1 + + ■ * ■ + (7.14)
where each is a polynomial of degree j and all polynomials, such as and
are orthogonal to each other. The advantage of writing the model in this
form is that additional polynomials of higher degree may be added which are
independent of the ones already considered. The values of these polynomials
are given for the first five powers of u as
«0 = 1
=
= A-
/3k^ - T
= A. U' u{ - (7.15)
20
.2
= A. u
n
= A, U' (k^ - 7) + (15/c* - 230k^ + 407)
18 1008
where k is the number of levels of the factor X, and the A’s are chosen so that
the are integers for all u. Because of the complications of these formulas
and the practical need to test for significant high-order regression models,
the values of the and A’s have been tabled by Fisher and Yates [11], and
a section is reproduced as Appendix Table F.
The A/s are given by
L.J Yjjij (7.16)
' L.J a?"
To see how these formulas and tables apply, consider a quadratic model
with just three points on the X scale. The u/s would be — 1, 0, + 1. As these
are already integers, A^ = 1 and takes on values —1, 0, +1 for each j.
For {'2, {k^ — 1)/12 = (9 — 1)/12 = 2/3 and the values are -l-l, 0, +1,
which makes [t/^ — (/c^ — 1)/12] ^ 1/3, —2/3, 1/3 so A2 is taken as equal
to 3 in order to make the ^'2 integers. Then the {'2 are +1, — 2, +1. Appendix
Table F gives these values for /c = 3. Also given in the table are ^ (fj)^ for
each set of and the Aj. In the expression for Aj, each Y^j is multiplied by
the corresponding {j, and since — 0. this expression is a contrast. Since
the coefficients — 1, 0, +1 are orthogonal to 1, —2, 1, these provide two
orthogonal contrasts on the }//s (or their totals, T j) and the sum of squares
7.4 Orthogonal polynomials 141
cc (Zi S; Yu ir
^ ^
" ZJ iQ
One needs only to apply the coefficients given in Appendix Table F to the
treatment totals of a single-factor problem to determine the sum of squares
for linear, quadratic, cubic, quartic, and so on, effects up to the (k — l)th-
degree equation. Equations (7.14), (7.15), and (7.16) are only needed if the
prediction model = /(u, u^, . . ., u^~^) is desired.
To illustrate this, consider the example of Sections 7.2 and 7.3. Since
k = 5 dosage levels. Table F gives the following coefficients for the linear,
quadratic, cubic, and quartic effects.
T, 83 84 89 96 115 I
j
ii’/ Q Contrast SS
Linear -2 -1 0 1 2 10 1 76 192.53
Quadratic 2 -1 -2 -1 2 14 1 38 34.38
Cubic -1 2 0 -2 1 10 5/6 8 2.13
Quartic 1 -4 6 -4 1 70 35/12 12 0.69
Total 229.73
Applying each set of coefficients to the treatment totals gives the contrasts
as listed and their corresponding sum of squares. For example, the cubic
contrast is
-1(83) 2(84) + 0(89) - 2(96) + 1(115) = 8
This expression is very similar to the expression for the sum of squares of a
contrast, except that the contrast is not squared and it will retain its sign.
142 Qualitative and quantitative factors
Source df SS MS F Significance
Totals 14 249.73
0 AQ = y. = 31.13 4'o = 1
76
1 2.53 = iw
“ 3(10) ■
38
2 A2 0.90 i'2 = - 2)
“ 3(14) “
8
3 A3 0.27 <^3 = 5/6(M^ — 3Au)
“ 3(10) "
12
4 •44 0.06 i'4 = 35/12
“ 3(70) “
TL = Aoi'o + A,S',
= 31.13 + 2.53u
as given in Section 12.
For a quadratic
FL = A,S'o + + ^2C2
- 31.13 + 2.53u + 0.90(u^ - 2)
which simplifies to
which simplifies to
“Fitting” this quartic polynomial gives Table 7.8. This table shows the
quartic to go through the five means except for slight rounding errors.
Uj Y'
T3
T.2
TA
X
Xi X3
A problem will now be considered where one factor is set at qualitative levels
and the other at quantitative (and equispaced) levels.
Example 7.2 We wish to detenuine the effect of both depth and position
in a tank (Figure 7.3) on the concentration of a cleaning solution in ounces
per gallon. Concentrations are measured at three depths from the surface
of the tank, 0 inch, 15 inches, and 30 inches. At each depth measurements
+ + +CREATE PROB8
1. 000 = 77509, 113
2. 000 = COMMON (SPSS)
3. 000 = SPSS
4. 000 = #EOR
5. 000 = RUN NAME STAT 502 PROB8
6. 000 = VARIABLE LIST RT, DOS
7. 000 = INPUT FORMAT (F2.0, IX, Fl.O)
8. 000 = N OF CASES 15
9. 000 = ONEWAY RT BY DOS (1, 5)/
10. 000 = POLYNOMIAL - 4/
11. 000 = STATISTICS ALL
12. 000 = READ INPUT DATA
13. 000 = 26 1
14. 000 = 28 1
15. 000 = 29 1
16. 000 = 28 2
17. 000 = 26 2
18. 000 = 30 2
19. 000 = 28 3
20. 000 = 30 3
21. 000 = 31 3
22. 000 = 32 4
23. 000 = 33 4
24. 000 = 31 4
25. 000 = 38 5
26. 000 = 39 5
27. 000 = 38 5
28. 000 = FINISH
O o
o
o ON
o v-H
o O'!
r^i
o <N
NO •»-H
o o O i-H <N
DH o o o o ITN m
[JH
o
HH r- o o GN m
H i-H NO o ON o NO
< r- (N (N ■1 1 P ro
od NO NO r-'
CNI ON
[JU
GO
W
cti
< rn m o o tT) m O
m m o 1—t ON fO GN o
:D
m m o oo O m oo o
a Nn m NO p
GO r-’ c4 (N r4 r4
Ln ON fO
z
pj
u “
Z S
<
M—I
Cei
< w
> CcJ
UH <
O Z) m
m
m o
m o
O
1—1
O
ON
m
ro GN
O
o
m
m
GO C/ ro m o OO <■ H m OO o m
GO
GO GN <N m p NO o p
IJH
ON <>i r4 (N o ON
<N ON ro m (N
<
O <N (N
z
0-)
< D
(y:i
’a
Sa
x tZ m <N
W
Q
,o
=5
O
< Q u
HH
*n
DH GO
w
z < 03
HH D
< P
P a U GO
> PJ D p P
O U O w p P S Z)
z Od P H o w
H
o p o o
< D O P p p H p p
O
GO z <
U
p
Q
p U
HH
p
o
w
w
h-^
z >■ < > m > ;z; <
r- w D p D p HH
H
2 H
p Q a Q U Q z
H
O
HH H
H
146
7.7 Two factors—one qualitative, one quantitative 147
are taken at five dififerent lateral positions in the tank. These are considered
as five qualitative positions, although probably some orientation measure
might be made on them. At each depth and position, two observations are
taken. This is then a 5 x 3 factorial with two replications per cell (total of
30 observations). The design is a completely randomized design and the
model is
^ijk = + Di Pj -{■ DPij +
with
z = 1, 2, 3 7 = 1, 2,..., 5 k = 1,2
where D, represents the depth and Pj the position. The data collected are
shown in Table 7.11.
In this case where one factor is set at quantitative levels (Dj) and the
other at qualitative levels (Pj), the first step is to run a two-factor analysis
of variance just as if both were qualitative factors. Coding the data of Table
7.11 by subtracting 5.90 and multiplying by 100 gives Table 7.12.
(-44)-
= 50.47
30
(1)" + (1)" + (4)" + ■■• + (-11)" (-44)-
SS„ X P interaction
30
- 281.67 - 50.47 = 58.33
SSerror = 579.47 - 281.67 - 50.47 - 58.33 = 189.00
giving the ANOVA table of Table 7.13.
Table 7.12 Coded Cleaning-Solution Concentration Data with Totals
1 0 0 4
1 -1 -10
1 -1 -6 -6
2 0 -1 -15
1 -1 -7
1 -2 -22 -23
3 4 1 -4
0 1 -7
4 2 -11 -5
4 3 4 -7
1 0 -1
4 4 -8 0
5 0 4 -7
-3 0 -4
-3 4 -11 -10
Ti.. 7 7 -58
II
1
i4
k=l j=l
37 37 570 L
i
44
J k
= 644
7.7 Two factors—one qualitative, one quantitative 149
Source df SS MS
Totals 29 579.47
From Table 7.13 it is seen that only depth produced a significant effect
on concentration of solution as
140.83
11.24
12.60
(-65)^
SS linear 211.25
10(2)
(-65)^ 70.42
SS quadratic
10(6) 281.67
Z (& A
Linear -1 0 +1 2 1
Quadratic +1 -2 +1 6 3
2(6^ 1^ =
Note that the sum of these two sums of squares (41.50 + 16.83) equals
the D X P interaction sum of squares (58.33) as it should. Summarizing for
this problem with one quantitative and one qualitative factor, the complete
ANOVA breakdown is as shown in Table 7.15.
The results in Table 7.15 indicate a strong depth effect, with the linear
depth effect significant at the 1 percent level (**) and the quadratic depth
effect significant at the 5 percent level (*). There is no position nor interaction
between depth and position. These results seem reasonable from a graph of
cell totals versus depth and positions (Figure 7.4).
This graph shows little interaction, as the curves are quite parallel
(statistically speaking) and there is little difference between the five position
curves. The depth effect is quite obvious, and concentration is seen to drop
7.7 Two factors—one qualitative, one quantitative 151
Source df SS MS
Depths 2 281.67
Linear 1 211.25 211.25**
Quadratic 1 70.42 70.42*
Positions 4 50.47 12.62
D X P interaction 8 58.33
l>linear X P 4 41.50 10.38
n
^quadratic
Y p
^ 4 16.83 4.21
Error 15 189.00 12.60
Totals 29 579.47
+5
Cell
Totals
(Coded)
-10
-15
-20
0 15 30
Depth (inches)
off with increasing depth but on more of a curve than a straight line. Further
investigation is suggested at depths between those already studied. The lack
of any position effect or interaction should mean that this new experiment
could be run at only one position; the results should then be the same at all
five positions.
If one wishes the equation for predicting concentration (coded) from
depth,
X - 15
u =
15
and
Ao= Y.= -44/30 - -1.47
A, = -65/10(2) - -3.25
A2 = -65/10(6) - -1.08
152 Qualitative and quantitative factors
and
- 1
S'l =
Xj Uj Y'
^ U
Y
0 -1 0.70 0.70
15 0 0.69 0.70
30 1 -5.80 -5.80
Lacquer Concentration
Standing
1
Time 2 1 li 2
30 16 12 17 13
14 11 19 11
20 15 14 15 12
15 17 18 14
10 10 7 10 9
9 6 14 13
7.8 Two factors—both quantitative 153
i = 1, 2, 3, 4 j = 1, 2, 3 k= 1,2
and the analysis can be shown to be as given in Table 7.17. These results
indicate strong main effects—lacquer concentration and standing time—but
no significant interaction. One could then extract linear, quadratic, and
cubic effects for the lacquer concentrations and linear and quadratic effects
for the standing times as two independent factors. However, since some
examples may have significant quantitative by quantitative interactions, we
will go through the steps to partition the interaction as well as the main
effects into component parts. For lacquer concentrations there are four
levels, or /c = 4 in Table F. To apply these coefficients one needs the four
lacquer totals. These are seen, along with other useful totals, in Table 7.18
based on Table 7.16.
Source df SS MS F Probability
Totals 23 262.95
Lacquer Concentration
Standing
Times j 1 l4 2 Tj
154 Qualitative and quantitative factors
T. 79 67 93 72 ii'/ Contrast SS
Linear -3 -1 1 3 20 2 5 0.21
Quadratic 1 -1 -1 1 4 1 -9 3.37
Cubic -1 3 -3 1 20 10/3 -85 60.21
Total = 63.79
Linear -1 0 1 2 1 35 76.56
Quadratic 1 -2 1 6 3 -49 50.02
Total SSj = 126.58
These results partition the main effects into components with 1 df each and
they can be tested for significance. To break down the interaction term, note
that there are six single-degree-of-freedom terms
LL X LQ X TLQ X
LL
-3 -1 1 3
-3 -1 1 3
1
30 23 36 24
0 0 0 0
TL 0
30 31 33 26
3 1 -1 -3
-1
19 13 24 22
7.8 Two factors—both quantitative 155
-3-11 3
-3 -1 1 3
1
30 23 36 24
6 2 -2 -6
TQ -2
30 31 33 26
-3 -1 1 3
1
19 13 24 22
^^LQXTL ~~ 5.06
^^LQ.TQ= 4.69
^^LcxTL ^ 2.81
TQ =
12.60
156 Qualitative and quantitative factors
Source df SS MS F Probability
Lacquer 3 63.79
Linear 1 0.21 0.21 <1 0.791
Quadratic 1 3.37 3.37 1.17 0.301
Cubic 1 60.21 60.21 20.91 0.001
Time 2 126.58
Linear 1 76.56 76.56 26.58 0.000
Quadratic 1 50.02 50.02 17.37 0.001
L X T interaction 6 38.07
FL X U 1 7.81 7.81 2.71 0.125
FL TQ 1 5.10 5.10 1.77 0.208
1 5.06 5.06 1.77 0.208
LQ X TQ 1 4.69 4.69 1.59 0.231
Fc X U 1 2.81 2.81 <1 0.343
Fc X TQ 1 12.60 12.60 4.38 0.058
Error 12 34.51 2.88
Totals 23 262.95
Adding the six components gives 38.07, the interaction sum of squares
reported in Table 7.17. The complete breakdown and tests of significance
can be summarized as in Table 7.21. This table shows three significant effects:
a cubic effect of lacquer concentration on screen quality, and a linear and
quadratic effect of standing time on screen quality. Since there is no signifi-
cant interaction one might plot the mean responses to each main effect as
in Figure 7.5.
Because the two main effects do not interact one might find an equation
to predict the lacquer means (a cubic) and another equation to predict
standing time means (a quadratic). These two equations could be combined
to predict the cell or treatment means. Use Equation 7.15 and the contrasts
found in the data to give
For lacquer (/c = 4):
^/s
= F = 311/24 = 12.9583
A, = 5/6(20) = 0.0417
.4 2 = -9/6(4) = -0.3750
A, = -85/6(20) = -0.7083
and
7' = 12.9583 + 0.0834U - 0.375(u^ - 1.25) - 0.7083(10/3)(u=' - 2.05M)
7.8 Two factors—both quantitative 157
Figure 7.5 Screen quality averages versus lacquer concentration and standing times.
which simplifies to
^0 =: Y 12.9583 io = 1
•4i —
35/8(2) = 2.1875 = iM
A, -49/8(6) = -1.0208 ^2 = 3(M" - 2/3)
and
Y' = 12.9583 + 2.1875U -- 1.0208(3u^ -- 2)
or
Y'
^ u - 14.9999 + 2.1875t/ -- 3.0624M^
Now the lacquer concentration equation will have u values as coded for the
lacquer data and the standing time equation for coded standing times.
Calling the lacquer variable
- 1.25
0.5
158 Qualitative and quantitative factors
To see how good a predictor this is. Table 7.22 gives the predicted
value at the upper left of each cell using this equation and the observed
mean of the cell at the lower right of that cell. If one now examines the de-
partures of the cell mean in Table 7.22 from their predicted values, the sum
of squares of the departures is
(15.0 - 14.33)2 + (11.5 - 12.33)2 + • • • + (11.0 - 8.79)2 _
Since each mean represents two observations
7.9 SUMMARY
The examples of this chapter may easily be extended to higher order fac-
torials whenever one or more factors are considered at quantitative levels.
The use of orthogonal polynomials makes the analysis rather simple, pro-
vided the experiment is designed with equispaced quantitative levels. The
summary of designs at the end of Chapter 6 has not been changed by this
chapter, as these methods can be used in all experiments where quantitative
levels are involved.
Chapter 8 presents a more general method for handling many inde-
pendent variables in determining their effect on some dependent variable 7.
PROBLEMS
7.1 For the following data on X and Y plot a scattergram and determine the least
squares straight line of best fit.
A
3 4 5 6 7
7 8 10 11 10
7 8 8 9 9 10
9 9 9 10 9
7.2 For Problem 7.1 present your results in an ANOVA table and comment on these
results.
7.3 From your table in Problem 7.2 find and the standard error of estimate.
7.4 Use the method of orthogonal polynomials on the data of Problem 7.1 and find
the best fitting polynomial.
7.5 An experiment to determine the effect of planting rate in thousands of plants per
acre on yield in bushels per acre of corn gave the following results.
Plot a scattergram of these data and comment on what degree polynomial might
prove appropriate for predicting yield from planting rate.
7.6 From the data of Problem 7.5 use orthogonal polynomials and determine what
degree polynomial is appropriate here.
7.7 Find the equation of the polynomial in Problem 7.6.
7.8 From an ANOVA table of the results of Problem 7.6 find R^, and comment
on what these statistics tell you.
7.9 Seasonal indexes of hog prices taken in two different years for hve months gave
the following results.
Month
Year 1 2 3 4 5
a. Sketch a scatterplot of these data and comment on what type of curve might
be used to predict hog prices when the month is known.
b. Do a complete analysis of these data and find the equation of the best fitting
curve for predicting hog price from month.
7.10 A bakery is interested in the effect of six baking temperatures on the quality of its
cakes. The bakers agree on a measure of cake quality, Y. A large batch of cake mix
is prepared and 36 cakes poured. The baking temperature is assigned at random
to the cakes such that six cakes are baked at each temperature. Results are as
follows:
Temperature
o
0
oc
22 4 10 22 32 20
10 21 22 14 27 36
18 18 18 21 22 33
26 28 32 25 37 33
21 21 28 26 27 20
21 28 25 25 31 25
Do a complete analysis of these data and write the ‘best’ fitting model for predicting
cake quality from temperature.
7.11 Compute statistics that will indicate how good your fit is in Problem 7.10 and
comment.
Problems 161
Source df SS MS
Total 39 98.675
a. Determine the significance of each prediction equation and state which degree
equation (linear, quadratic, or cubic) will adequately predict retention score
from age.
b. Write the equation for part (a).
c. Find the departure from regression of the mean retention score of a 37-year-old
based on your answer to part (b).
7.13 A study of the effect of chronological age on history achievement scores gave the
following results.
History Achievement
Chronological Age
O .. _ J
T - 18
J 1- -
ocol t/ L)y
^ 15 8 9 10 11 12
5 1 6 5
4 1 7 4 5
3 6 2
2 3 2
1 6 1
0 1
7.14 In studying the effect of the distance a road sign stands from the edge of the road
on the amount by which a driver swerves from the edge of the road, hve distances
were chosen. Then the five distances were set at random, observing four cars at
each set distance. The results gave
rj^ = 0.70
= 0.61
162 Qualitative and quantitative factors
a. Test whether or not a linear function will “ht” these data. (You may assume a
total SS of 10,000 if you wish, but this is not necessary.)
b. In fitting the proper curve, what are you assuming about the four readings at
each specified distance?
7.15 Thickness of a film is studied to determine the effect of two types of resin and three
gate settings on the thickness. Results showed:
Resin Type 2 4 6
Resin Type
1 2
Gate Weight Fraction Weight Fraction
Setting
(mm) 0.20 0.25 0.30 0.20 0.25 0.30
7.22 Data on the effect of knife edge radius R in inches and feedroll force F in pounds
per inch on the energy necessary to cut 1-inch lengths of alfalfa are given as follows:
5 29 98 44 84
30 128 81 100
20 67 77 63
79 293 202 247 821
10 22 35 53 103
26 80 93 90
16 29 59 98
64 144 205 291 704
15 18 49 58 80
17 68 103 91
11 61 128 77
46 178 289 248 761
20 38 68 87 86
31 74 116 113
21 47 90 81
90 189 293 280 852
Write the mathematical model for this experiment and do an ANOVA on the
two factors and their interaction.
7.23 Outline a further ANOVA based on Problem 7.22 with single-degree-of-freedom
terms.
7.24 Do an analysis for each of the terms in Problem 7.23.
7.25 Plot some results of Problem 7.22 and try to argue that they confirm some results
found in Problem 7.24.
8
Multiple Regression
8.1 INTRODUCTION
In Chapter 7 we considered a prediction model to predict some dependent
variable Y from one independent variable X where the model may be some
polynomial in X. We also considered examples of two or more independent,
quantitative variables for cases where the levels of each such variable were
set at equispaced values. We now consider a more general situation in which
Y may be a function of several independent variables X2,..., with
no restrictions on the settings of these k independent variables. In fact, in
most such multiple regression situations the X variables have already acted
and we simply record their values along with the dependent variable Y. This
is, of course, ex-post-facto research as opposed to experimental research
where one manipulates the X’s and observes the effect on Y.
In practice there are many studies of this type. For example, one may
wish to predict the surface finish of steel from drop-out temperature and
back-zone temperature. Here 7 is a function of two recorded temperatures
X^ and X2. It may be of interest to predict college grade-point average
(GPA) in the freshman year for students whose input data include rank in
high school, high school regents’ average, SAT (Scholastic Aptitude Test)
verbal score, and SAT mathematics score. Here T, the freshman year GPA,
is to be predicted from four independent variables: X^—high school rank,
X2—high school regents’ average, X3—SAT verbal, and X^—SAT mathe-
matical.
As in the previous chapter, a mathematical model is written and the
coefficients in the model are determined from the observed sample data by
the method of least squares, making the sum of squares of deviations from
this model a minimum.
164
8.2 The multiple regression model 165
To predict the value of a dependent variable (7) from its regression on several
independent variables (X^, X2,..., X^) the linear population model is
given as
r = /So + + ■■■ + /SA
where the ^’s are the true coefficients to be used to weight the observed X’s.
In practice a sample of N is chosen and all variables (7, X2,..., X;^)
are recorded for each of the N items in the sample and the corresponding
sample model is
where the B's are estimators of the j5’s. (Capital B's are used here because
computer programs do not usually provide for lowercase letters.) Note that
if /c = 1, Equation (8.1) gives
Y' = Bo Y B,X,
7, = 7; + i= 12,..., N
by making the sum of the squares of the errors of estimate (^ ef) a minimum,
one can derive the following least squares equations.
y y = BoW + I -^1 + 1 ^2 + • ■ ■ + B. z
y x, y = Bo I + Bi I + B y
2 + ... + y x,x,
y x,y = Bo y X, + Bi y + B, y + ■ ■ ■ + B,y x^x, (8.2)
y x,Y = Bo y X* + Bi y x,x, + B, y + ■ ■ ■ + B, y 2^,^
This set of /c + 1 equations in /c + 1 unknowns {Bo, B^, . . ., 5 J can be solved
for the B's. The solution is tedious if more than two or three independent
variables are involved and so these equations are usually solved by a com-
puter program that can handle a large number of independent variables.
To see how such problems are handled consider a problem from Burr
[8],
Example 8.1 Twenty-four samples of a steel alloy are checked as to their
percent elongation under stress and for each sample a chemical analysis is
made and the percentages of five specific chemical elements are recorded.
166 Multiple regression
These are identified only as X^, X , X , X , and X^. The data are given
2 3 4
in Table 8.1.
To enter Equations (8.2) the following sums are computed from Table
8.1.
N = 24
yx, = 11.68 y X? = 5.9038 y XI262 = 35.553
1^2 = 73.7 'ZXl = 240.85 y 261X3 = 10.101
1363 = 22.1 yxl = 31.17 y X1X4 = 47.644
y 264 = 96.1 y = 408.09 y X1X5 = 0.18405
y X5 = 0.397 Y^Xl = 0.0145 y XiY = 67.318
y y = 136.2 ^ = 943.78 y X2X3 = 71.17
y ^2X4 = 297.57 y X3X4 = 74.58 y X4X5 = 1.2881
y XjX; = 1.2886 y 2('32C5 = 0.5945 X ^47 = 563.35
y Xjy = 390.32 y XjF = 102.53 y XjY = 1.6384
8.2 The multiple regression model 167
ANOVA df SS MS F Significanee
The value of R SQUARE indicates that 51.314 percent of the variation (SS)
in Y is accounted for by the above equation. Note: Total SS = 170.845 and
87.66795/170.845 = 0.51314. The F statistic is significant at the 0.016 level,
indicating the equation accounts for a significant part of the total variation.
The STD DEVIATION reported is really the standard error of estimate—
the standard deviation of the errors remaining after the prediction equation
has been “fitted” to the data. The standard deviation of Y prior to prediction
is 2.73:
I - (Z ^943.78
(136.2)2
N 2^
2.73
N 23
If R2 — 0.51314 of the variation in Y is accounted for by the regression of
Y on Xi, X2, . . ., X5, then 1 — = 0.48686 is unaccounted for or residual.
So the SS of the error or residual should be
SS error SSy(l 12345)
170.845(0.48686)
83.177
168 Multiple regression
and
SS error '83.177
12345 2.15
N - k - 1 18
as given.
Sometimes the multiple R is designated as 12345 to show the joint
effect of all five X's on Y. In like manner, Sy 12345 denotes the standard
error of estimate in Y from the regression model based on five X's.
The question might now be raised as to whether or not one needs all
five independent variables to predict Y. If a Stepwise Regression program is
used, the computer will enter the X's in the order of their contribution to
the prediction of Y. At each step in the process the input can be examined
to see whether or not adding another variable will contribute significantly to
the prediction of Y. If not, one may end up with a reduced regression equa-
tion based on a subset of the X’s that will be adequate in predicting Y.
Our data were entered into an SPSS Stepwise Regression as shown in
Table 8.2.
The printout will be considered in several stages to see what one can
determine from the computer. The first part of the printout gives the mean
and standard deviation of each variable.
Y 5.6750 2.7254 24
X, .4867 .0977 24
X2 3.0708 .7948 24
A3
.9208 .6859 24
A4
4.0042 1.0063 24
As .0165 .0186 24
Ai .16884
A2 -.56054 -.17600
A3 -.53234 -.42456 .26356
A4 .28508 .38712 .13389 -.87641
As -.52747 -.21924 .20448 .78077 -.70099
Y A, A, A3 A4
In examining the correlations between Y and each X, one notes that the
highest numerical r is T with X2 {ry2 = —0.56054) and thus a Stepwise Re-
gression will enter X2 first into the regression equation and give the resulting
simple linear equation and other useful information as seen in Table 8.3.
UJ
8.2 The multiple regression model 169
+ + + CREATE PROB4
1.000 = 77509,113
2.000 = COMMON (SPSS)
3.000 = SPSS
4.000 = #EOR
5.000 = RUN NAME STAT 502 PROB4
6.000 = VARIABLE LIST Y, XI, X2, X3, X4, X5
7.000 = INPUT FORMAT (5F4.0, F5.0)
8.000 = N OF CASES 24
9.000 = REGRESSION METHOD = STEPWISE/
10.000 = VARIABLES = Y, XI, X2, X3, X4, X5/
11.000 = REGRESSION = Y WITH XI, X2,
X3, X4, X5/
12.000 = STATISTICS ALL
13.000 = READ INPUT DATA
14.000 = 11.3 .50 1.3 .4 3.4 .010
15.000 = 10.0 .47 1.2 .3 3.6 .012
16.000 = 9.8 .48 3.1 .7 4.3 .000
17.000 = 8.8 .54 2.6 .7 4.0 .022
18.000 = 7.8 .45 2.8 .7 4.2 .000
19.000 = 7.4 .41 3.2 .7 4.7 .000
20.000 = 6.7 .62 3.0 .6 4.7 .026
21.000 = 6.3 .53 4.1 .9 4.6 .035
22.000 = 6.3 .57 3.7 .8 4.6 .000
23.000 6.3 .67 2.7 .6 4.8 .013
24.000 = 6.0 .54 3.1 .7 4.2 .000
25.000 = 6.0 .42 3.1 .7 4.4 .000
26.000 = 5.8 .33 2.6 .6 4.7 .008
27.000 = 5.5 .51 3.9 .9 4.4 .000
28.000 = 5.5 .54 3.1 .7 4.2 .000
29.000 = 4.7 .48 4.0 1.1 3.7 .024
30.000 = 4.1 .38 3.3 .8 4.1 .000
31.000 = 4.1 .39 3.2 .7 4.6 .016
32.000 = 3.9 .60 2.9 .7 4.3 .025
33.000 = 3.5 .54 3.2 .7 4.9 .022
34.000 = 3.1 .33 2.9 2.9 1.0 .063
35.000 = 1.6 .40 3.2 3.2 1.0 .059
36.000 = 1.1 .64 2.5 .7 3.8 .018
37.000 = .6 .34 5.0 1.3 3.9 .044
38.000 = FINISH
39.000 = #s
PJ
u
z
<
u O
o
z
a
GO
>
H -H Ol
r- o
< m o
U m Tf
H o o ll-
PU H ^ —: OS
CO r-
CQ o
< 7 ' o
-J
PJ
PJ
u
z o so 00 o ITN so NO
PJ < 00
CNI
Os
so NO o o
ro
p m P
PJ DC u 00
U 00 NO OO
<
z oo 'sf ro o ON
PM NO ro
o
< o c?
Tf c5
o
o
oo
ON
NO
ITN z NO irS r-*
u ON • G^
a
CO r- <N Z o
r- ro NO m
rsi O
X z o
z •TN
►-H
H
2
o
h-^
z o 'NO
m <
PJ <
o
c3 H
< in a
PJ
u
Q a
PJ
z <N r- ON
PJ GO
PJ < O o
<N
1-^
00
c CO
PJ
X DC Os
SO
o
00
_o cC
PJ
DC DC
H PJ OS ON ON ON
-I—> w X O
Tf OO
m < z MJ
CQ H NO 1-^ 1:3 os O
OD DC OO H
c z DC
<N
<r)
ON
r-
T—<
O ON
r^
t/N
NO
H
PJ O a^
in
NO 1—1 O
W z CO
PJ Q
NO PM rO
T—<
z -J
CO ON UTi <N o
c Q. Tt O H O
< o 00 ro
W IT) (N O PQ CO PJ
o H
O Tf m
< MJ H
NO
OO 00
oo
m
OS
O
G in
>0 --I
in m
OO O
<N ro D
ca DC o m
_o fN
5 CO < <
t/D Z < ON
r^i
oo
oo 2 CP
t/D 0 > <
D r~
VH Q PJ DO
r^l
(N
r-
m H >
(=0 tu os U PJ
<u cC < pp CP MJ
pj Q r- CQ
D m
H
az <N o < Tf
X
in
Z c/0 O (N ■It
2 X X
OH UP
W 0£ Osi P H
<
>
H PJ
(>0 PJ w PJ Q j z PJ
>
HJ Oi PJ > oa < U H
oa
< PS|S
< H
CO
z
<
G
•-H 2
<
z CQ
o 01
< > <>l o 2
< <
u
> S Di < GO X > 2
<
PM
o z >
GO
o -J PM
cn in
CO
CO
< o
00 > PJ pp
MJ DC Q
-2
2i < O 2
pp
S
Z PJ PJ O
H < DC DC U
170
8.2 The multiple regression model 171
as the best fitting straight line for predicting Y from X2 alone. With each
B is recorded its standard error and a test as to whether it is significantly
greater than zero. Here both BQ and B2 are highly significant (probabilities
of 0.000 and 0.004 respectively).
This printout also shows an ANOVA for the single X’s effect of Y and
the multiple R and which, in this first step, is the simple Pearson r of
y versus X2. Thus this equation will account for 0.3142 of the variation in
y. Note from the ANOVA that the linear regression on X2 with 1 df has
an SS of 53.68 out of a total (53.68 + 117.17) of 170.85, or 53.68/170.85 -
0.3142. The STD DEVIATION is actually the standard error of estimate or
standard deviation of the errors of estimate from this first equation. Here
Sy 2 = 2.31, which is some reduction from the original Sy of 2.73.
The “Variables Not in the Equation” portion of the printout records
partial correlation coefficients. These are correlations between Y and each
X after X2 has been removed. They are sometimes designated as ry^ 2? and
so forth, indicating the correlation between y and X^ with X2 removed.
These four partials are now examined to see whether any are significantly
greater than zero. Here three of the four are (at the 0.05 level) and the largest
partial is Vys.j = —0.50903, so X5 is next entered into the regression equa-
tion as STEP 2 as printed out in Table 8.4.
The resulting equation is now
Y' = 11.69 - 1.62X2 - 63.I8X5
At this stage Ry 25 = 0.49209 and the standard error of estimate is 2.03. An
examination of the next set of partial r’s
^yl.25 = -0.01274
^y3.25 — —0.16275
r,4.25 = 0.09925
shows none significant at the 0.05 significance level. Hence one may stop in
this problem after X2 and X5 have been entered and there is no added
significance in going to an equation with the other three variables included.
One might also note that the SS for the error after X2 and X5 have
been accounted for is 86.77 whereas it is only 83.18 when all three X’s are
included. One might also note that the standard error of estimate appears to
UJ
U
z
<
p o
o
>
H o — r- NO
NO ON —'
<
H
U m
ONI
r-
NO
oo
NO
00
o —
(DO r-
oo
UJ H ro r-
CQ GO
<
-J
UJ UJ
U 04
z OI
UJ
UJ < NO r- ON o
UJ DC u rO lO O NO ON NO
u ON ON ON 00 NO
< UH wn OO
Z r- NO ro ro O P ON ^ 1— ON
< C^ o to 1—H NO
OO o m P a
m
iT)
—
(N z OJ
OO
ON
"TD
Z
o
u
<N
00 O
ON
CO m
p
m
— Z g ro
U. to oo (N Tf g
X p ro z
H
< z od r-' <
UJ
H
<
o
■<c—<3
P p UJ
a CO o
UJ U
(N UJ
z
Q UJ
GO
UJ <
00
ON ^
ON Tt
c X CQ UJ X DC ro r-
500
(N
o CC H
DC DC H UJ ON
03
UJ O
ON
o
to
<N
1—H
<N < J
o
CJ) ca DC NO rn p
GO ON to ON r- (N H
c DC r- C^ ON a o ^ H
UJ (N 00 CO r-
-J UJ p O
p r-
s Z CQ Q
rn
(N u. Tit \6 z
c cu ON ON (N
< H O 00 00 CO 'Ct r- (N
o UJ o r-- r-"
ro <N 2
GO UJ
HJ H
04
>0
ON
ON
c
H
GO O
r- ■ TT O
ON ro < P CQ DC P o
o H
> GO < <
• Z oo Ot NO l-H f
C/D tn NO (N OH
(/) o ON 00 DC
<u ON ON <
ON r- ON
&0
Q UJ oa NO H >
UJ CC p u UJ
<u Qi
< NO 0- P
0:5 UJ m
Q
<N H P oq m Tt
OH
Z
UJ
az
00 O
CNI
CNl m
—GO
QC
P 'X
X X
W DC DC H f-H <
>
\
H UJ
CO UJ UJ UJ Q < J z UJ
p DC UJ > < >
m U
CQ p < H UJ < H H
CO CO Z
< H P Q
2 <
3 J o P Z
Qi < oa
o < p CO Q Q
> CNI tn o DC
<
H u <
> DC < GO X X > 2
iH
UM <
OH
O >
CO -J UH
h-N CO
CO
< o
00 >- UJ p u.
J DC Q UH
2 < o ^ S
es z UJ UJ O
CtC
H < I U
111
8.2 The multiple regression model 173
One might also note in this table that the residuals or errors of estimate
seem to be nonrandom as in general they are positive for large T’s and
negative for small T’s. This may indicate a need to consider some higher
order terms such as X\, X\, or X X^. These could be entered into the
2
T Y' e T Y' e
a: ^
tu
csi
< O
(N o
<dN
’-^
Tj- (N O
—H o^ O —I
w a m ^ CO
CO
HJ
OQ QC
i <
H
> Qi
CeS
W
< J ^ 0^ —I Tf
D- CO O CD r<o
h-H O ^ CO
H CO O —' —'
CO 1^
D J
I CO D
w
u
z
< ro oc Oc oo
u
1—1 O ^ cO CO CO
UH
p p cq
h-^
Z
o
HH
CO
UJ
>
O
yj —H CO O CXD
OO ON
O Oc O oc
H [V
CO Tj- o oo
p ro CO (N p
UH O O K ' ' ’
03
0<
w
Q
H
c
_o Z
■t—< w
o3
tJD
d
o Q
s w
a>
3
al
o3
H <
5 Q
c3 < w
a > Cii fN| CO ro
tu
a X z z z z
d H
CO Z
yj
'^O
00
OH
w Cxi OO CO
«s H
H CO
174
8.3 Computer programs 175
To test whether any more variables are needed after a subset r has been
found significant the test is
Rl - Rf/k - r
(8.4)
1 - Ri/N - fc - 1
where represents the correlation of all k values with Y and the correla-
tion of a subset r of the k variables with Y.
In our example = 0.51314 and R^ = 0.49209. Testing for signifi-
cance of all five variables:
0.51314/5
3.79
0.48686/18
as in ANOVA. Testing whether one needs to go beyond X2 and X^,
_ 0.51314 - 0.49203/3
“ 0.48686/18 ^
hence nonsignificant.
Many computer programs give a SUMMARY TABLE at the end of
their printout, which shows the variables entered in order of their entry and
the multiple R and R^ increase as more variables added will account for more
and more of the variability in the Y variable. In this example Table 8.6 gives
this information. One notes that after adding X5, 49 percent of the variation
in Y is accounted for and after that additional variables increase this only
slightly. The increase can be tested for significance at any stage by Equation
(8.4).
more variables can be removed and all those left are contributing significantly
to the prediction equation.
A FORWARD method will enter the variables one at a time until the
equation is found such that entering any more variables will not improve
the prediction significantly. Again, one may designate the size of F necessary
for addition to the equation.
A STEPWISE method is similar to the FORWARD method except that
at each stage all variables already entered are examined to see whether they
are all needed after entering the last variable. In some cases an earlier entry
is dropped. For example, if X^, and X7 were already selected in that
order, when X2, the next most significant variable, is entered X^ might be
removed. This may be unnecessary, now that X2 has been added, and the
equation will include X4, X7, and X2 only.
In all these methods default values are in the programs that will operate
in adding or removing variables if one does not instruct the computer as to
the level desired. It is also possible to include certain specified X’s in the
equation regardless of their significance.
In the problems at the end of this chapter it is assumed that some
multiple regression program is available to the reader.
8.4 REMARKS
and Zj =
If this is done, the resulting equation for predicting the standardized 7 can
be shown to be given by
Here
= or =
Sy Sj
and the rs are simple Pearson r’s. Solving these equations for the B'^s is still
difficult for very many independent variables. In the special case where /c = 1,
i3 y
B\ = Vyi and B^ = — •
(8.6)
and
Bo=Y B,X,
In Example 8.1 the first variable entered was Xj and the preliminary data
gave the following.
Y = 5.6750 Sy = 2.7254
Xj = 3.0708 S2 = 0.7948
and
^Y2 - -0.56054
Hence
Sy 2.7254
(-0.56054) = -1.92
^ S2 0.7948
and
Bo = Y - B2X2 = 5.6750 - (-1.92)(3.0708) = 11.58
PROBLEMS
8.1 The sales volume E of a product in thousands of dollars, price Xj per unit in
dollars, and advertising expense X2 in hundreds of dollars were recorded for n =
178 Multiple regression
Case
1 2 3 4 5 6 7 8
a. Find the means and standard deviations of each variable and the intercorre-
lation matrix.
b. On the basis of the discussion in Section 8.4 find the regression equation Y
from XI alone.
8.2 Submit the data of Problem 8.1 to a Stepwise Regression program and find the
regression equation and the value of Ry 12 decide whether or not both X'’s
are necessary.
8.3 In studying the percent conversion (7) in a chemical process as a function of time
in hours and average fusion temperature in degrees Celsius {X2) the following
data were collected.
Y ^2
62.7 3 297.5
76.2 3 322.5
80.8 3 347.5
80.8 6 297.5
89.2 6 322.5
78.6 6 347.5
90.1 9 297.5
88.0 9 322.5
76.1 9 347.5
Analyze these data using multiple linear regression. Reexamine these data and
comment on any pecularities noted.
8.4 A management consulting firm attempted to predict annual salary of executives
from the executives’ years of experience (^1), years of education (^2), sex (X3), and
number of employees supervised (^^4). A sample of 25 executives gave an average
salary of $29,700 with a standard deviation of $1300. From a computer program
the following statistics were recorded.
= 0.42
RL4 = 0.78
R? 124 = 0.90
^r.i234 = 0.95
Problems 179
Use this information to test for the significance of each added variable and deter-
mine whether or not an ordered subset of these four will make a satisfactory
prediction. If you wish, you may assume a total of sum of squares in this study of
10,000 units. Also explain how you would predict with approximate 2 to 1 odds
the maximum GPA expected for an individual student based on his or her scores
of the variables in the appropriate equation.
8.7 In a study of several variables that might affect freshman GPA (T) a sample of
55 students reported their college entrance arithmetic test (Xj), their analogies
test score on an Ohio battery (X2), their high school average (X3), and their
interest score on an interest inventory (X4). Results of this study are given in the
following (oversimplified) ANOVA table.
Source* df SS MS
Total 54 4000
* / means “given.”
180 Multiple regression
Assuming a stepwise procedure was used so that variables were entered in their
order of importance, answer the following.
a. Determine RY.I234 explain its meaning in this problem.
b. Determine which of the four variables are sufficient to predict GPA (F) and
show that your choice is sufficient.
c. If a regression equation based on your answer to (b) is used to predict student
A’s GPA based on test scores, how close do you think this prediction will be
to the actual GPA? Justify your answer and note any assumptions you are
making.
8.8 In an attempt to predict the average value per acre of farm land in Iowa in 1920
the following data were collected.
Y = average value in dollars per acre
Xi = average corn yield in bushels per acre for ten preceding years
Xi = percentage of farm land in small grain
X3 = percentage of farm land in corn
Y ^2 -Y3
87 40 11 14
133 36 13 30
174 34 19 30
385 41 33 39
363 39 25 33
274 42 23 34
235 40 22 37
104 31 9 20
141 36 13 27
208 34 17 40
115 30 18 19
271 40 23 31
163 37 14 25
193 41 13 28
203 38 24 31
279 38 31 35
179 24 16 26
244 45 19 34
165 34 20 30
257 40 30 38
252 41 22 35
280 42 21 41
167 35 16 23
168 33 18 24
115 36 18 21
Analyze these data using multiple linear regression and comment on the results.
Problems 181
8.9 In a study involving handicapped students an attempt was made to predict GPA
from two demographic variables, sex and ethnicity, and two independent measured
variables, interview seore and contact hours used in counseling and/or tutoring.
A dummy variable for sex was taken as 1 = male, 2 = female. For ethnicity:
1 = black, 2 = Hispanic, and 3 = white, non-Hispanic. Do a complete analysis
of the data below and justify the regression equation that will adequately “fit”
the data.
8.10 Given the following intereorrelation matrix between measures where Y = fresh-
man mathematics grade, = Ohio State psychological examination, X2 =
English-usage examination, A = algebra examination, A = engineering apti-
3 4
tude test:
182 Multiple regression
Variable Y ^2 ^3 ^4
0.51
^2 0.51 0.70
^3 0.61 0.53 0.61
^4 0.39 0.39 0.29 0.28
Means 5.70 4.10 5.44 5.37 4.95
Standard deviation 2.42 1.92 1.84 2.26 2.14
X 403 = ^9^10
X 404 = ^9^12
405 = ^10^13
This means, standard deviations, and intercorrelations are given on p. 183. (Note
that Y is X^.)
a. On the basis of the above printout what is the hrst variable to be entered in a
stepwise regression? Explain how you know.
b. On the basis of just the information given hnd the best fitting prediction
equation for predicting ^5 from this first variable.
c. How good is this prediction?
d. What is the standard error of estimate?
8.12 At step 4 in Problem 8.11 the printout is as appears on p. 184.
Explain what the next step should be and discuss the results up to this point.
That is, what is the equation and how good is it at step 4?
ON <N ON <r) 1-H
o NO (N O
OO ON
1'-' 00 rn r- {y^
m cn m ro <;
CASES
oooooooooooooooo >
rNi<NrNirNirsi<Nr^(N
o T-H m T-H o
NO i-H o
OO 00
m 1-H 1-H (N
CT'i o O o
<
STANDARD DEV
1 1 >'
00 so o NO* r- oo o o O 80 m
T—1
00 OO so so ON OS NO
m so oo ro oo ON OO CNI O
r- r- p p (N r- oo 1-H 00 (N (N
NO p (N (N (N
<
r-‘ rn oo os od
O O o >'
o(N
CN so r- O
NO •,..1 r-
OO OO r- OS oo 1
r-4 o •rs o o (N o
o rn 1-H ro ^-H ro 0^
ro o ON OO ON ro
Q <
PJ >
so Os vzs H
i—H SO (N ro •rs OO to to 00 ON ro NO ro
MEAN
rn OO OS so so OS ro NO ro 00 CD ON t-H t-H
OS p SO p so OO (N OH o o 00 NO ON 80 ON r-' O
NO 0^4 r- r- r- (N ivl
O rxi (N so o S ro o 04 ro ro oo
r--
oo
SO ro
Tf 1-^ 00
t—H so (N
o 1
<
>
u
(N m
w r- r- NO ON NO NO NO OO to (N
CQ oo t-H ro r- ON r- ON ■^H
r- Tt to ^H to ro O
H rg NO to 04 00 ro t-H 80 NO fy^
VARIABLE
rN o ro
•ri' uri o (N ro
o o o o o o
o CM to 04
<r'
O O ■rf- ■Tf z
z
< O ro OO ro ro O ro
< < < < < < < < o OS 8 00 oo SO o ro SO 00
> > > > > > > > u NO o NO ro ON SO ON r- 0
Ovj oo ON ON to OO NO o (N to 01
H CM ro (N ro ro r--
z <
>
CASES
PJ
oo OO 00 00 OO oo OO OO
<N (N CNI CN^ (N (N (N u OO 80 r- 00 80 to 80
Tf SO ro o ro ro SO 80 o
UH os to r- so OS to r- r- oo 0
Lin NO (N r- Tf ON to ON oo t-H ro 01
ro os <N o ro ON 00 ON ro
PU <
o 1 >
u
STANDARD DEV
o (N r- r- ON NO ro r- ro T—H ro oo
< r- to 80 ON 80 o 04 to oo O
00 ON ON ON OO ON
ro ro 00 os os OS OO SO 0"^ to ro (N O
pp to r- 00 80 o Tf ro r- 80 OO ro SO Cxi
00 00 O SO ro ro ro (N NO ro ON ro ro ro ro
rn OO —H Tt so <
r- ON NO r- m o
NO rn NO (N rN
Q
PJ
>
NO 80
m H T-H r- ofN o <N
VAR404
(N NO 00 80 NO
.29343
ON
1-^ z r- <N (N 80 O o ON 04 ro t-H
r- oo OO —H to NO to 00 ■t-H o t-H OO
o NO oo to 1—H oo ON r- 1-H r- 00 to ro ex:
ro r- o ro r- o ro r- NO r- (N
f~ CL <
z ^ 1 >
S ^
Qs o o ro NO ro ro —-H ro to rs)
.57640
.79807
U o
VAR403
(N o r- <N oo OO -H O O ro (N oo ro r- so m ro G^. ro 00 OO o
rn ro so o fO ON so P^ O to 00 oo SO so ro ON ro to os ro 0
to to so 00 SO r- ON 80 to
MEAN
OS oo ro (N O so pp s p (N r- o 'Tf (N r- O CN so r-
01
so •rs
o6 H
oo O
CM oo
ro
oi to
S ^ ■
1
<
<N
a-.
80 m
r-
80
o
o^ >
T—^ to r- OO o so r- so i-H ON ro ro o oo NO rg
.34442
VAR402
lo
.88608
.92208
Z o o r-- ON NO
ON to Tf NO
ON
NO
to
80
80 O NO (N
oo NO (N
ro ON ro 00 NO
OO to o ON
o
o
O PJ (N ON! o 80 80 to ro O ro 1-H oo 00 O D£;
OO OO ro ON ro ON (N O ro OS r- oo ro
H ^ <
>
VARIABLE
NO OO ON o CNJ ro
-J <
o o o
rs o o PJ >• r- OO ON o ONI ro 80 o (N ro
VAR404
o so to
VAR403
yAR405
u > > > > > > > > > > > > > > >
183
w o w
o
u o u
Z Z so—'^0<N^r~-Os(Nroosro OOIOO—'■^^sO-^sO'-O
< < r-^ooOsOO^O^O'/'iOOro-^rf-osoorOooOO^OsO
r''U-iroO'^'^r~'(N—i-Hr'--™soosoo>oornrsioorN|
u u so • NO • o • oo • CNJ * NO • <N
h-H [-!-( ON <N y—t NO ■»—( (N r- r-- OO Os
UH IJH r- 00 NO NO Tf NO CN y—t
HH m wo CNj NO r4
o o p p ON.
z Z z NO r-; r4 r4 r^i r^i
o O o
So H GO
<
Z)
lO a PJ
ro w O
o^ z
ro UJ m m oo NO ro o ro NO
O X < (N
OO
<N r-
T—(
rn
(N
O OO
r-j
Ov
r-
00 CM
■—
r<-l Di O 00 O
H i-H NT) (N r^i wo ro o r^i
Csl PU o o OO r- oo wo 00 o (N OO
z HJ
O
PJ OO IT) H H
oi O <N
r-- t"- 0
< o o z HJ lO)
o ON NO 04 o- CN 04
(Jr oo
O GO < o OS 1—( wo O ON ON 00
PJ 04 ro O i-H ON ro o- ro
03 CM J H ro NO NO O- ON 00 OO NO wo
OS CQ Cci O 04 o o O o o o o o o
CNl
Z < <
•< rn CP
W 01
<
> PP
-J r-- OS o Ol m G-) o n-) G-)
GO 00 m o o i-H ^H ^H o o O o o
m ro OO o o O o o o
oo < DC DC DC DC DC DC DC DC DC DC DC
0£ r^i so
0£ <c < < < < < < < < < <
< Tj- \o < > > > > > > > > > > >
IZ) oo r-
t" so
>
5/5 so oo
OS <N
(j_ '
^ m
U so
>-
H NO NO O- 00 WO ■_<
00 ro 0- O oo ON
< NO 04 r- wo o r- WO
GO y~-l ro T-^ ro ON o o-
GIFT AMOUNT-CURRENT YEAR
r
P-^ Os p NO 04 y—t ON o
P-1 (
■»— ■»—( WO oo
PQ GO r- 04 o 1 o
H <
U _1
OH PP
so
PH m pp
<N
Q U
Z ON O wo O 04 o r^) oo NO
VAR402
PP T—1 ro o 04 ro
u >- < 04 o WO P NO
H u NO i-H
z HH
J
PH ON
O
wo G^
04
ro
^—(
r-
< PH
a< O ro
Z wo O ro y^
S 3
< o z Osl ’—1 (N
<
>
HH HH
H
a
l-H
DC 00
<
4 . .
<
o z
PH
> D
GO
o HP pp
a
pp
cri
So
00 GO <C o pp
o
VARIABLE(S) ENTERED ON STEP NUMBER
> PP w PP
HP DC Q
PH
X DC o O
PH
H l-H oo ON ON O
C a So PP
z
DC NO ON 04
z PP PP O pp
l-H
04 wo
Ovl
NO
ON
o o
< DC DC U
HH OO r- 04 ro GO
VAR005
GO
Q l-H ro 04 ro no’
H o6 04 no
PP GO
^ G-l ^ O J
SO OO (N CQ
OS O O so
GO Csi —H <
HH
os Os os no O
DC
so PP
oo < NO o ON
DEPENDENT VARIABLE. .
no > NO
ON
NO NO o
04
ro
ro Os) ro 04
r- r- r- O
PP 04 NO 04 ^H
DC O) wo wo NO
r- oo
< NO
ro
00 "=:f
WO
a Z
GO O
DC DC H H
PP pp Q < pp
Z
HP DC pp HP
CP
> QQ <
H PP < so oo CM H
GO
H D Q o
o o O GO
HP a D DC o o Z
DC DC DC DC
D c/3 Q Q
H
<
> < < < < o
S DC < c/3 > > > > u
184
p
U
o
z
< P
w o r \ r- GO r- fN o o (N GO o m CO (N r~ 0-1 ON
o U (N oo OO (N GO o oo OS o 1 4 0-1 <N T-H
u o OO (N T—H ■ r -H (N ro O (N oo GO GO T-H
[J-N GO T—H o OI • t" p ON O
z (N 0-1 oo
< o z ON
GO
GT) so
oo
rn
GS
ON
oo
o
oo
o GO
0-1
ON
m
O
oo
u O ON
H o (N t"; p P p p
l-u < CO
D
z a
a w
p
U
00
PJ z
< v-H
r-
T-H
OO
OA
OS
yn
o
ro
(N GO
0-1
o
(N
r-
X on 1■ \ OO
H O GO T-H
(N m ON (N o ON) o
UJ p oo. r" p P P oo.
m z ■] C P
O
u. r-; H H
lA o
oo
oo z P (N ON on 0-1 ON OO oo
GO < O (N ro >o O m OO oo o
HH r- t" ON O m m
PJ H o ro lo 0^ 0-1 GO r^i P
p p p p p P p P p p
p o
p (N < l' l’
r- < p
< GO
p GO S
od <N < p
a (N > p r- O (N ro o ^—1 ro 0-1
COm O p O ^—1 o O O O O
ON ON o O O O P
Z fO <, P P
p P P P p P P
< m p < < < < < < < < <
P < > > > > > > > > >
GIFT AMOUNT-CURRENT YEAR
>
Ui 00 OO
nj H ■^r'rNir''0(N'0GD<^r<-)r<l0
O CNI O Ot^OOCNroC-'^CNl^O^OO
0^ o o^ < r<^'^^00t^r~-0^0^00O'^O
< p P U OGD'^(Nun'Oror''rnir)’^r<-i
H HH
(N H ■
D D OO P 00
^ • m ,■ ON
CTl a ^ m P c o pic C I P I
ON <
p
VAR015
o p
o\ oo
m ^n
p
U
z 0-1 O r<-) O r-
oo 0-1
-H O ON o
r~- ■^H ON ON ■^H O
1 T-H T
D- p o
OO o 0-1 OO p
OS o (N GO O o o-i
U- ^ (N o
H
<d
z <N ^—1 rn m ro NO
Q o
D
VARIABLE(S) ENTERED ON STEP NUMBER
a
P- p
p
O H p
gj O p p p o o
GO X
VAR005
U 00 <
00
O
m
H o P
CNI O GO
P
0-1 Csl o
>' Z W D p p OO ON rn oo ON T—H
< oi Q p < Z
I—( p 0-1 o
<N
oo
ON
ON
CNl
i-H
o
T-H
ON
< oi O 00 p P GO ON) o oo
p p 00 0-1 OO p
z < w p O < p Q
o p
o-i
p
rd
< > os: P u > p H p od rn P
p 00
DEPENDENT VARIABLE . .
< ro ON ON NO
GO 04
o O 0-1 5 r- (N ON ON O OO
GO GO 0-1 OO 00 a\ m 0-1 m rn
<N GO 0-1 ro < rn oo r-- oo T-H GO
GO
ON
(N
ON
(N
ON
r^i
os
> P
oo
oo
r<'i
t";
T-H
p
GO 0-1
0-1
T-H
p
p
o T—H P m ON OS GO
r- OO (N P r— r- (N
m
P P
P p Q
z p H
z
o p
P P P l-H p <
P
< H P H < GO OO (N ON 0-1 H
H p 00 < HH o o O o 00
P a D<
HH oi o o O o o z
P CO Q a Q > < p p P P p p
o
H P > < < < < <
u
P < GO 00 Q > > > > > >
185
186 Multiple regression
Explain the next step and analyze the results as completely as you can.
8.14 In a study on the effect of several variables on posttest achievement scores in
biology in a rural high school involving 48 students the following table was
presented. It included variables that had a significant effect on achievement at
the 10 percent level of significance.
Multiple R^
Step Variable Entered R R^ Change
Discuss these results and test whether or not the last two variables were necessary.
8.15 To study the effect of planting rate on yield in an agricultural problem, 13 data
points were taken with the following results.
V 12 12 16 16 16 20 20
Y 130.5 129.6 142.5 140.3 143.4 144.8 144.1
X 23 23 23 23 28 28
Y 145.1 147.8 146.6 148.4 134.8 135.1
a. Plot a scattergram and note what type of curve you might expect of Y versus X.
b. Taking = X2 and X = X^, use multiple regression to find the equation
of the best fitting quadratic.
c. Compare this quadratic with the best linear fit.
8.16 The printout of a computer program yielded the following statistics.
9.1 INTRODUCTION
As 2^ factorial experiments represent an interesting special case of factorial
experimentation, so also do 3^ factorial experiments. The 3^ factorials
consider / factors at three levels; thus there are 2 df between the levels of
each of these factors. If the three levels are quantitative and equispaced, the
methods of Chapter 7 may be used to extract linear and quadratic effects
and to test these for significance. The 3^ factorials also play an important
role in more complicated design problems, which will be discussed in
subsequent chapters. For this chapter, it will be assumed that the design is a
completely randomized design and that the levels of the factors considered
are fixed levels. Such levels may be either qualitative or quantitative.
9.2 3^ FACTORIAL
If just two factors are crossed in an experiment and each of the two is set at
three levels, there are 3x3 = 9 treatment combinations. Since each factor
is at three levels, the notation of Chapter 6 will no longer suffice. There is now
a low, intermediate, and high level for each factor, which may be designated
as 0, 1, and 2. A model for this arrangement would be
187
188 3^ Factorial experiments
2r • • •
02 12 22
• • •
Factor B 01 11 21
0 • • •
00 10 20
0 1
Factor A
first digit indicates the level of factor A, and the second digit the level of
factor B. Thus 12 means A at its intermediate level and B at its highest level.
This notation can easily be extended to more factors and as many levels as
are necessary. It could have been used for 2^ factorials as 00, 10, 01, and 11,
corresponding respectively to (1), a, b, and ab. The only reason for not using
this digital notation on 2^ factorials is that so much of the literature includes
this (1), u, h, . . . notation. By proper choice of coefficients on these treatment
combinations, the linear and quadratic effects of both A and B can be
determined, as well as their interactions, such as A^ x /l^ x BQ, AQ X
Bi^, and AQ X BQ. The methods of analysis will be illustrated in a simple
hypothetical example.
Example 9.1 Suppose the responses in Table 9.1 were recorded for the
treatment combinations indicated in the upper-left-hand corner of each cell.
Factor A
Factor B 0 1 2 T.i
00 10 20
0 1 -2 3 2
01 11 21
1 0 4 1 5
02 12 22
2 2 -1 2 3
T. 3 1 6 10
9.2 3^ Factorial 189
Source df SS MS
Ai 2 4.22 2.11
Bj 2 1.56 0.78
AB,j 4 23.11 5.77
Totals 8 28.89
Analyzing the data of Table 9.1 by the general methods of Chapter 5 gives
(10)"
SStotal — l" + O" + 2^ + • • • + 2^ = 28.89
ss. ^ ^ 4.22
,, _ 2^ + 5^ + 3^ (10)^ _ ,
SSg — ^ ^ — 1.56
Treatment Combination
Factor 00 01 02 10 11 12 20 21 22 Zcf
AL -1 -1 -1 0 0 0 +1 +1 +1 6
AQ +1 +1 +1 -2 -2 -2 +1 +1 +1 18
BL -1 0 +1 -1 0 +1 -1 0 +1 6
BQ +1 -2 +1 +1 -2 +1 +1 ~2 +1 18
ALBL +1 0 -1 0 0 0 -1 0 +1 4
ALBQ -1 +2 -1 0 0 0 +1 -2 +1 12
AQB]^ -1 0 +1 +2 0 -2 -1 0 +1 12
AQBQ +1 -2 +1 -2 +4 -2 +1 -2 +1 36
Yu 1 0 2 -2 4 -1 3 1 2
190 3^ Factorial experiments
From Table 9.3 it can be seen that Ai^ compares all highest levels of
/4(+1) with all lowest levels of A{—1). AQ compares the extreme levels with
twice the intermediate levels. Both of these effects are taken across all levels
of B. Now compares the highest versus the lowest level of B at the 0 level
of A, then at level 1 of A, then at level 2 of A, reading from left to right across
Similarly, BQ compares the extreme levels of B with twice the intermediate
level at all three levels of A. The coefficients for interaction are found by
multiplying corresponding main-effect coefficients. An examination of these
coefficients in the light of what interactions there are should make the
coefficients seem quite plausible. The sums of squares of the coefficients are
given at the right of Table 9.3.
Applying these coefficients to the responses for each treatment com-
bination gives
A^ = -1(1) - 1(0) - 1(2) + 0(-2) + 0(4)
+ 0(-l) + 1(3) + 1(1) 1(2) - 3
AQ = +1(1) + 1(0) + 1(2) - 2(-2) - 2(4)
- 2(-l) + 1(3) + 1(1) + 1(2) = 7
B^ = -1(1) + 0(0) + 1(2) - l(-2) + 0(4)
+ 1(-1) - 1(3) + 0(1) + 1(2) = 1
BQ = +1(1) - 2(0) + 1(2) + l(-2) - 2(4)
+ 1(-1) + 1(3) - 2(1) + 1(2) = -5
A^B^ = +1(1) + 0(0) - 1(2) + 0(-2) + 0(4)
+ 0(-l) - 1(3) + 0(1) + 1(2) = -2
A^BQ = -1(1) + 2(0) - 1(2) + 0(-2) + 0(4)
+ 0(-l) + 1(3) - 2(1) + 1(2) = 0
AQB^ = -1(1) + 0(0) + 1(2) + 2(-2) + 0(4)
- 2(-l) - 1(3) + 0(1) + 1(2) ^ -2
AQBQ = +1(1) - 2(0) + 1(2) - 2(-2) + 4(4)
- 2(-l) + 1(3) - 2(1) + 1(2) = 28
The corresponding sums of squares become
SS,,,, = ^ = 0.00
SSB, = ^ = 0.17
Source df SS
A, 2 4.22
AL 1 1.50
AQ 1 2.72
2 1.56
BL 1 0.17
BQ 1 1.39
AB,^ 4 23.11
ALBL 1 1.00
ALBQ 1 0.00
AQBI^ 1 0.33
AQBQ 1 21.78
Totals 8 28.89
/I
Plotting AI^BQ means using all three levels of B, giving Figure 9.3, which
again shows little interaction.
For AQBI^ the results are as shown in Figure 9.4, which once again shows
little interaction. This graph does indicate that the quadratic effect of A is
more pronounced than its linear effect. This is borne out by the data.
Replotting Figure 9.3 with B as abscissa gives Figure 9.5, which shows
the same lack of serious interaction as in Figure 9.3, but does show a slight
quadratic bend in factor B over the linear effect.
So far no startling results have been seen. Now plot the middle level of
B on Figure 9.4 to obtain Figure 9.6. Note the way this “curve” reverses its
trend compared with the other two. This shows that not until both factors
are considered at all three of their levels does this AQBQ interaction show up.
It can also be seen by adding the middle level of A on Figure 9.5 (see Figure
9.7).
9.2 3^ Factorial 193
BQ
Bi
Bx
Factor A Factor A
Factor B 0 1 2
0 1 -2 3 1 -2 3
1 0 4 1 0 4 1
2 2 -1 2 2 -1 2
Before leaving this problem, reconsider the data of Table 9.1. Add the
data by diagonals rather than by rows or columns. First consider the diag-
onals downward from left to right where the main diagonal is 1 + 4 + 2 = 7,
the next one to the right is —2 + 1 + 2= 1, and the last +2 is found by
repeating the table again to the right of the present one as in Table 9.5.
Similarly, the next downward diagonal gives 3 + 0 + (—1) = 2. The
sum of squares between these three diagonal terms is then
If the diagonals are now considered downward and to the left, their
totals are
3 + 4 + 2 = 9
1 + 1 - 1 = 1
-2 + 0 + 2 = 0
and their sum of squares is
These two somewhat artificial sums of squares of 6.89 and 16.22 are seen to
add up to the interaction sum of squares
I{AB) - 6.89 2 df
J{AB) = 16.22 2 df
total/I X B = 23.11 4 df
9.2 3^ Factorial 195
Each such component carries 2 df. These are sometimes referred to as the
AB^ and AB components of A x B interaction. In this notation, effects can
be multiplied together using a modulus of 3, since this is a 3^ factorial. A
modulus of 3 means that the resultant number is equal to the remainder when
the number in the usual base of 10 is divided by 3. Thus 4 = 1 in modulus 3,
as 1 is the remainder when 4 is divided by 3. The following associations also
hold:
numbers: 0 3 = 0 6 = 0 9 = 0
1 4=1 7 = 1 10 = 1
2 5 = 2 8 = 2 11 =2
When using the form A^B"^, it is postulated that the only exponent allowed
on the first letter in the expression is a 1. To make it a 1, the expression can
be squared and reduced, modulus 3. For example,
A^B = {A^Bf = A^B^ = AB^
Hence AB and AB^ are the only components of the A x B interaction with
2 df each. Here the two types of notation are related as follows:
I{AB) = AB^
J{AB) = AB
To summarize this simple experiment, all effects can be expressed with
2 df each, as in Table 9.6.
Source df SS
At 2 4.22
Bj 2 1.56
1{AB) = AB^ 2 6.89
J{AB) = AB 2 16.22
Totals 8 28.89
It will be found very useful to break such an experiment down into 2-df
effects when more complex designs are considered. Here, this breakdown is
presented merely to show another way to partition the interaction effect.
Let us consider a practical example.
Example 9.2 The effect of two factors—prechamber volume ratio {V) and
injection timing (T)—on the parts per million of noxious gas emitted from
an engine was to be studied. The volume ratio was set at three equispaced
196 3^ Factorial experiments
Volume {V)
Time
(T) Low Medium High
levels and the time was also set at three equispaced levels. Two engines
were built for each of the nine treatment combinations and the amount of
gas emitted was recorded for the 18 engines as shown in Table 9.7.
The ANOVA for Table 9.7, assuming a completely randomized design,
is given in Table 9.8.
Source df SS MS F Prob.
The ANOVA of Table 9.8 shows strong main effects and an interaction
significant at the 5 percent significance level. As both factors are quantitative
and equispaced we can treat these data as we did the hypothetical data of
Table 9.3 to give Table 9.9.
Upon taking each contrast and squaring and dividing by 2 (as n = 2)
and the sum of the squares of the coefficients, one gets Table 9.10.
From Table 9.10 we see that volume ratio produced both a strong
linear and quadratic effect on the amount of gas emitted. The timing shows
a strong linear effect and a significant interaction is shown as a quadratic
by quadratic interaction. One must then plot all nine treatment means to
see the significant effects as shown in Figure 9.8.
Note that without the T2 line in Figure 9.8 there would be little obvious
interaction. Careful study of this illustration should confirm the reason-
ableness of the results given in Table 9.10.
9.2 3^ Factorial 197
Factor 00 01 02 10 11 12 20 21 22 Contrast
Tu. 11.70 13.45 14.27 16.12 15.00 17.62 15.21 16.93 18.09
Source df SS MS F Prob.
V- 2 11.44 5.72
VL 1 9.74 9.74 121.75*** 0.000
''Q 1 1.70 1.70 21.25** 0.001
Tj 2 4.17 2.08
TL 1 4.03 4.03 50.38*** 0.000
TQ 1 0.14 0.14 1.75 0.219
4 1.38 0.35
VLTL 1 0.01 0.01 <1
VLTQ 1 0.00 0.00 <1
VQTL 1 0.25 0.25 3.12 0.111
1 1.12 1.12 14.00** 0.005
Error 9 0.70 0.08
p = 2 3 4 5 6 7 8 9
3.20 3.95 4.42 4.76 5.02 5.24 5.43 5.60
and LSRs = 0.64, 0.79, 0.88, 0.95, 1.00, 1.05, 1.09, 1.12.
From these the lines can be drawn as shown above. Clearly, the mini-
mum emission is when one has low prechamber volume ratio and the
shortest injection time as the mean of 5.85 ppm is significantly lower than
any other of the nine means.
9.3 3^ FACTORIAL
If an experimenter has three factors, each at three levels, or a 3 x 3 x 3 = 3^
factorial, there are several ways to break down the effects of factors A, B,
and C and their associated interactions. If the order of experimentation is
completely randomized, the model for such an experiment is
Factor A
Factor B Factor C 0 1 2
1 0 010 no 210
1 on 111 211
2 012 112 212
Source df Source df
A 2 AQCJ^ 1
AL 1 AQCQ 1
AQ 1 BCj, 4
Bj 2 BLCL 1
BL 1 BLCQ 1
BQ 1 BQCL 1
ABij 4 BQCQ 1
ALBL 1 ABC,J, 8
ALBQ 1 ALBLCL 1
AQBJ^ 1 A^B^CQ 1
AQBQ 1 ALBQCJ^ 1
Q 2 ALBQCQ 1
Q 1 AQBJ^CJ^ 1
CQ 1 AQBJ^CQ 1
AQ, 4 AQBQC 1
ALCL 1 AQBQCQ 1
ALCQ 1
Total 26
200 3^ Factorial experiments
Source df
A 2
B 2
AB 2)ia
AB^ 2) r
C 2
AC 21la
AC^ 2| r
BC 21
la
BC^ 2r
ABC 21
ABC^ 2
^8
AB^C 2
AB^C^ 2j
Total 26
Example 9.3 A problem involving the effect of three factors, each at three
levels, was proposed by Professor Burr of Purdue University. Here the
measured variable was yield and the factors that might affect this response
were days, operators, and concentrations of solvent. Three days, three op-
erators, and three concentrations were chosen. Days and operators were
qualitative effects, concentrations were quantitative and set at 0.5, 1.0, 2.0.
Although these are not equispaced, the logarithms of these three levels are
equispaced, and the logarithms can then be used if a curve fitting is warranted.
For the purposes of this chapter, all levels of all factors will be considered
as fixed and the design will be considered as completely randomized. It was
decided to take three replications of each of the 3^ = 27 treatment com-
binations. The data, after coding by subtracting 20.0, are as presented in
Table 9.14.
9.3 3^ Factorial 201
Day Di
5/14 5/15 5/16
Operator Oj
Concentration A B C ABC A B C
0.5 1.0 0.2 0.2 1.0 1.0 1.2 1.7 0.2 0.5
1.2 0.5 0.0 0.0 0.0 0.0 1.2 0.7 1.0
1.7 0.7 -0.3 0.5 0.0 0.5 1.2 1.0 1.7
1.0 5.0 3.2 3.5 4.0 3.2 3.7 4.5 3.7 3.7
4.7 3.7 3.5 3.5 3.0 4.0 5.0 4.0 4.5
4.2 3.5 3.2 3.5 4.0 4.2 4.7 4.2 3.7
2.0 7.5 6.0 7.2 6.5 5.2 7.0 6.7 7.5 6.2
6.5 6.2 6.5 6.0 5.7 6.7 7.5 6.0 6.5
7.7 6.2 6.7 6.2 6.5 6.8 7.0 6.0 7.0
Source df SS MS
Di 2 3.48 1.74**
Oj 2 6.14 3.07**
DO,j 4 4.07 1.02**
Q 2 468.99 234.49***
4 0.59 0.15
OCjk 4 0.89 0.22
8 1.09 0.14
54 9.98 0.18
Totals 80 495.23
202 3^ Factorial experiments
(a) {b)
Concentration Concentration
Totals 18.6 105.6 177.5 301.7 Totals 18.6 105.6 177.5 301.7
Sums of Squares
- -1(18.6) + 0(105.6) + 1(177.5) - 158.9
SSc,
CL = 27(2) = 467.58
The D X CL is then
Source df SS MS
Di 2 3.48 1 74**
Oj 2 6.14 3.07**
DO, 4 4.07 1.02**
CL 1 467.58 467.58***
1 1.41 2 41**
CQ
D X CL 2 0.49 0.24
D X CQ 2 0.09 0.04
0 X CL 2 0.70 0.35
0 X CQ 2 0.19 0.09
DOC,j, 8 1.09 0.14
^m{ijk) 54 9.98 0.18
Totals 80 495.22
This second analysis shows that the linear effect and the quadratic effect
of concentration are extremely significant. Two plots of Figure 9.8 may help
in picturing what is really happening in this experiment.
Figure 9.9(a) shows the effect of operators, days, and D x O interaction.
Figure 9.9(b) indicates that the linear effect of concentration far outweighs
the quadratic effect and there is no significant interaction. If a straight line
or three straight lines were fit to these data, the logs of the concentrations
would be used, as the logs are equispaced.
204 3-^ Factorial experiments
Although this would usually conclude the analysis of this problem, each
interaction will be broken down into its diagonal, or / and J, components
in order to illustrate the technique. To compute the two diagonal compo-
nents of the two-factor interactions, the two parts of Table 9.16 can be used,
along with a similar table for the D x O cells (see Table 9.18).
From Table 9.18 the diagonal components of the D x O interaction are
I{D X O)
(301.7)'
= 1.74
81
Call it DO\
= 2.33
Operator
Day A B C
C, D, at Cl D , at C2 D , at C3
Call it DO. These total 1.74 + 2.33 = 4.07, the D x O interaction sum of
squares.
Applying the same technique to the two parts of Table 9.16 gives
Now form a table with these I and J components at each level of C (see
Table 9.20).
l(DO) J{DO)
c, h h C jo 7i ji
Treat each half of Table 9.20 as a simple interaction and compute the
/ and J components. Thus,
(101.3)" + (98.6)" + (101.8)" (301.7)"
DO^C^ = /[C X /(DO)] = 0.22
27 81
(99.1)" + (99.4)" + (103.2)" (301.7)"
DO^C = J[C X /(DO)] = 0.39
27 81
(100.0)" + (102.8)" + (98.9)" (301.7)"
DOC^ = /[C X J(DO)] = - 0.30
27 81
(99.8)" + (102.4)" + (99.5)" (301.7)"
= 0.19
27 81
total D X 0 X C
compared with 1.09 in Table 9.17. This last breakdown into four parts could
also have been accomplished by considering the C x 0 interaction at three
levels of Di, or the C x D interaction at three levels of Oj.
The resulting analysis is summarized in Table 9.21. This analysis is in
substantial agreement with Tables 9.15 and 9.17. No new tests would be
performed on the data in Table 9.21, as they represent only an arbitrary
breakdown of the interactions into 2-df components. The purpose of such a
breakdown is discussed in subsequent chapters. For testing hypotheses on
interaction, these components are added together again.
Source df SS
D. 2 3.48
Oj 2 6.14
DO 2 2.33
DO^ 2 1.74
c,
DC
2
2
468.99
0.29
DC^ 2 0.30
OC 2 0.65
OC^ 2 0.24
DOC 2 0.19
DOC^ 2 0.30
DO^C 2 0.39
DO^C^ 2 0.22
^m(i jk) 54 9.98
Totals 80 495.24
Problems 207
9.4 SUMMARY
The summary at the end of Chapter 6 may now be extended for Part II.
+ ABij + £k{ij)^ ■ ■ ■
PROBLEMS
9.1 Pull-off force in pounds on glued boxes at three temperatures and three humidities
with two observations per treatment combination in a completely randomized
experiment gives:
Temperature A
9.4 Develop a “Yates method” for this 3^ experiment and check the results with those
above.
9.5 When the data were collected in Example 9.2 on ppm of noxious gas, data were
also recorded on percent smoke emitted by the 18 engines. Results were as follows:
Volume (F)
Time
(T) Low Medium High
inches, and 6 inches, and three subbase thicknesses; 4 inches, 8 inches, and 12
inches. Two observations were made under each of the 27 pavement conditions
and complete randomization performed. The results were as shown on page 208.
Do a complete ANOVA of this experiment by the methods of Chapter 5.
9.12 Since all three factors are quantitative and equispaced, determine linear and
quadratie effects for each factor and all interaction breakdowns. Test for signif-
icance in Problem 9.11.
9.13 Break down the interactions of Problem 9.11 into 2-df eomponents such as AB,
AB^, ABC, AB^C, and so on.
9.14 Use a Yates method to solve Problem 9.11 and check the results.
9.15 Plot any significant results of Problem 9.11.
9.16 From Example 9.3 fit the proper degree regression equation for the quantitative
variable and express it in terms of the original variable.
9.17 The following data are on the wet-film thickness (in mils) of lacquer. The factors
studied were: type of resin (two types), gate-blade setting in mils (three settings),
and weight fraction of nonvolatile material in the lacquer (three fractions).
Resin Type
1 2
Do an ANOVA of the above data and pull out any quantitative effects in terms of
their components.
9.18 Using a = 0.01, write prediction equations where possible and check them
(Problem 9.17).
9.19 If the gate setting is 5 mils and the weight fraetion 0.20, what is the best prediction
of wet-film thickness? If interaction is ignored, what is the prediction? How much
difference does it make? (Problem 9.17.)
9.20 Plot results indicated in Problem 9.18.
10
Fixed, Random, and Mixed Models
10.1 INTRODUCTION
In Chapter 1 it was pointed out that, in the planning stages of an experiment,
the experimenter must decide whether the levels of factors considered are
to be set at fixed values or are to be chosen at random from many possible
levels. In the intervening chapters it has always been assumed that the factor
levels were fixed. In practice it may be desirable to choose the levels of some
factors at random, depending on the objectives of the experiment. Are the
results to be judged for these levels alone or are they to be extended to more
levels of which those in the experiment are but a random sample? In the
case of some factors such as temperature, time, or pressure, it is usually
desirable to pick fixed levels, often near the extremes and at some inter-
mediate points, because a random choice might not cover the range in which
the experimenter is interested. In such cases of fixed quantitative levels, we
often feel safe in interpolating between the fixed levels chosen. Other factors
such as operators, days, or batches may often be only a small sample of all
possible operators, days, or batches. In such cases the particular operator,
day, or batch may not be very important but only whether or not operators,
days, or batches in general increase the variability of the experiment.
It is not reasonable to decide after the data have been collected whether
the levels are to be considered fixed or random. This decision must be made
prior to the running of the experiment, and if random levels are to be used,
they must be chosen from all possible levels by a random process. In the
case of random levels, it will be assumed that the levels are chosen from an
infinite population of possible levels. Bennett and Franklin [4] discuss a
case in which the levels chosen are from a finite set of possible levels.
When all levels are fixed, the mathematical model of the experiment is
called a fixed model. When all levels are chosen at random, the model is
called a random model. When several factors are involved, some at fixed levels
and others at random levels, the model is called a mixed model.
210
10.2 Single-factor models 211
Whether the treatment levels are fixed or random, it is assumed in this model
that /r is a fixed constant and the errors are normally and independently
distributed with a zero mean and the same variance, that is, Sij are NID
(0, <jj). The decision as to whether the levels of the treatments are fixed or
random will affect the assumptions about the treatment term TJ. The different
assumptions and other differences will be compared in parallel columns.
^3
^2
/ 1
1 1 1
M2 M Ml M3
(a)
Figure 10.1 Assumed means in (a) fixed and (b) random models.
EMS: 3. EMS:
A
k - 1 (7s + A k - 1 cTg -I-
i= 1
B/s are fixed B/s are NID (0, al) B/s are NID (0, ai)
constants and
i B, = 0
i=i
I AB,J = 0
i t AB<j = 0
i
Z AB,, = 0
i AB,J # 0
j
(for A fixed, B
random)
2. Analysis: 2. Analysis: Same 2. Analysis: Same
Procedures of
Chapter 5 for sums
of squares
3. EMS: 3. EMS: 3. EMS:
ab{n — 1) 'yl
214 Fixed, random, and mixed models
In the assumptions for the mixed model the fact that summing the
interaction term over the fixed factor is zero but summing it over the
random factor is not zero affects the expected mean squares, as seen
in item 3 on page 213.
For the fixed model the mean squares for A, B, and AB are each com-
pared with the error mean square to test the respective hypotheses, as should
be clear from an examination of the EMS column when the hypotheses are
true. For the random model the third hypothesis of no interaction is tested
by comparing the mean square for interaction with the mean square for
error, but the first and second hypotheses are each tested by comparing the
mean square for the main effect {A^ or Bj) with the mean square for the
interaction as seen by their expected mean square values. For a mixed model
the interaction hypothesis is tested by comparing the interaction mean
square with the error mean square. The random effect Bj is also tested by
comparing its mean square with the error mean square. The fixed effect (Ai),
however, is tested by comparing its mean square with the interaction mean
square.
From these observations on a two-factor experiment, the importance
of the EMS column is evident, as this column can be used to see how the
tests of hypotheses should be run. It is also important to note that these
EMS expressions can be determined prior to the running of the experiment.
This will indicate whether or not a good test of a hypothesis exists. In some
cases the proper test indicated by the EMS column will have insufficient
degrees of freedom to be sufficiently sensitive, in which case the investigator
might wish to change the experiment. This would involve such changes as a
choice of more levels of some factors, or changing from random to fixed
levels of some factors.
The two examples above have shown the importance of the EMS column
in determining what tests of significance are to be run after the analysis is
completed. Because of the importance of this EMS column in these and more
complex models, it is often useful to have some simple method of determining
10.4 EMS rules 215
these values from the model for the given experiment. A set of rules can be
stated that will determine the EMS column very rapidly, without recourse
to their derivation. The rules will be illustrated on the two-factor mixed
model of Section 10.3. To determine the EMS column for any model:
1. Write the variable terms in the model as row headings in a two-way
table.
^k{ij)
2. Write the subscripts in the model as column headings; over each sub-
script write F if the factor levels are fixed, R if random. Also write the
number of observations each subscript is to cover.
a b n
F R R
i j k
At
AB,j
3. For each row (each term in the model) copy the number of observations
under each subscript, providing the subscript does not appear in the
row heading.
a b n
F R R
i j k
Ai b n
a n
AB,j n
^k{ij)
4. For any bracketed subscripts in the model, place a 1 under those sub-
scripts that are inside the brackets.
216 Fixed, random, and mixed models
a b n
F R R
i j k
A, b n
a n
AB,j n
1 1
a b n
F R R
i j k
Ai 0 b n
a 1 n
AB,, 0 1 n
1 1 1
6. To find the expected mean square for any term in the model:
a. Cover the entries in the column (or columns) that contain non-
bracketed subscript letters in this term in the model (for example,
for Ai, cover column i; for cover column k).
b. Multiply the remaining numbers in each row. Each of these products
is the coefficient for its corresponding term in the model, provided
the subscript on the term is also a subscript on the term whose
expected mean square is being determined. The sum of these coeffi-
cients multiplied by the variance of their corresponding terms (0^ or
a^) is the EMS of the term being considered (for example, for T,-,
cover column i). The products of the remaining coefficients are bn, n,
n, and 1, but the first n is not used, as there is no / in its term {Bj).
The resulting EMS is then bncj)^ + -I- 1 • cr^. For all terms,
these rules give:
a b n
F R R
i J k EMS
d, 0 b n + najiB + nb(f)^
B, a 1 n + naal
/IB,, 0 1 u +
1 1 1
10.4 EMS rules 217
These results are seen to be in agreement with the EMS values for the
mixed model in Section 10.3. Here is, of course, a fixed type of variance
liAf
—
a
Although the rules seem rather involved, they become very easy to use with
a bit of practice. Two examples will illustrate the concept.
4 5 2
R F R
Source df i i k EMS
Ti 3 1 5 2 Cg + i0(Tj
Mj 4 4 0 2 CTg + 2(TJ]^ +
TM,j 12 1 0 2 CTg + 2<TYM
20 1 1 1
The proper F tests are quite obvious from Table 10.1 and all tests have
adequate degrees of freedom for a reasonable test.
6 4 5 2 6
R F F F R
Source df i j k m Q EMS
0, 5 1 4 5 2 6 + 240(TJ
-4; 3 6 0 5 2 6 + 60aoA + 3600^
0/1,; 15 1 0 5 2 6 (^e 60(7^A
C, 4 6 4 0 2 6 + 48(7OC + 2880c
20 1 4 0 2 6 + 48(7CC
12 6 0 0 2 6 + i2(7})AC + 720^c
OACij, 60 1 0 0 2 6 + 13(7’O4C
From this table it is easily seen that all interactions involving operators
and the operator main effect are tested against the error mean square at the
bottom of the table. All interactions and main effects involving fixed factors
are tested by the mean square just below them in the table.
The rules given in this section are general enough to be applied to the
most complex designs, as will be seen in later chapters.
Using the definitions of expected values given in Chapter 2 and the proce-
dures of Section 3.6 the EMS expressions may be derived for the two-factor
experiment.
For a two-factor experiment, the model is
with
/ = 1, 2,. . ., a j = I, 2, b k = 1,2, . . ., n
The sum of squares for factor A is
= SZj k
= ZZ j k
+ h(ij))/b^
+Z j
+Z j
+ Z Z ^kiij))/b^
j k
y... = Z Z Z yijk/abn
i j k
a b n
^ ZZZi j k
+ ^kiij))/^b^
F ,. = // + Z A,la + Z B^jb + Z Z AB.^/ab
I }
a b n
+ Z Z Z <^Hii)lnab
i j k
Subtracting gives
+ ( Zj Zk et(ij)/bn - Z Z Z enij)/abn i j k
Note that the B effect cancels out of the A sum of squares as it should, since
A and B are orthogonal effects in a factorial experiment. Squaring and
adding gives
SS^ = nb - Z AJa
i= 1 \ i
+ Z fz ABij/b - Z Z ABJab
j= 1 \ j i j
a /bn a b n \2
+ Z ZZ
i= 1 \ j k
- Z Z Z Haj)/abn \
i j k J
+ (cross-product terms)
220 Fixed, random, and mixed models
nb nb
£(SS^) = nb{a — l)aj + -^ (ab — b)G\B 3—yrr “ bn)G^
b n b
T(MS^) = nba^ + na^^ +
X /li = 0
i
t '45,7 = 0
i
but
Z AB,j 7^ 0
j
then
^ nb ub
£(SS^) - nb X [nab - nb)Gj
i b n b
a
which agrees with the value stated in Section 10.3 for a mixed model.
Using these expected value methods, one can derive all EMS values
given in Section 10.3. It might be noted that if A were random in the mixed
model
I AB,j = 0
j
then the interaction term would not appear in the factor A sum of squares.
This is true of B in the mixed model of Section 10.3.
These few derivations should be sufficient to show the general method
of derivation and to demonstrate the advantages of the simple rules in
Section 10.4 in determining these EMS values.
10.6 The pseudo-/’ test 221
Example 10.3 Two days in a given month were randomly selected in which
to run an experiment. Three operators were also selected at random from a
large pool of available operators. The experiment consisted of measuring the
dry-film thickness of varnish in mils for three different gate settings: 2, 4, and
6 mils. Two determinations were made by each operator each day and at
each of the three gate settings. Results are shown in Table 10.3.
Day
1 2
Assuming that days and operators are random effects, gate settings are
fixed, and the design is completely randomized, the analysis yields Table 10.4.
Source df SS MS EMS
Totals 35 1.7622
All F tests are clear from the EMS column except for the test on gate
setting, which is probably the most important factor in the experiment. Only
the three-way interaction shows significance. It is obvious from the results
that gate setting is the most important factor, but how can it be tested? If the
D X G interaction is assumed to be zero, then the gate setting can be tested
against the 0 x G interaction term. On the other hand, if the 0 x G inter-
action is assumed to be zero, gate setting can be tested against the D x G
interaction term. Although neither of these interactions is significant at the
5 percent level, both are numerically larger than the D x O x G interaction
against which they are tested. In this case any test on G is contingent upon
these tests on interaction. One method for testing hypotheses in such situa-
tions was developed by Satterthwaite and is given in Bennett and Franklin
[4, pp. 367-368].
The scheme consists of constructing a mean square as a linear combina-
tion of the mean squares in the experiment, where the EMS for this mean
squares includes the same terms as in the EMS of the term being tested, except
for the variance of that term. For example, to test the gate-setting effect G
in Table 10.4 a mean square is to be constructed whose expected value is
+ 2<TIOG + 4a^o +
An F test can now be constructed using the mean square for gate setting as
the numerator and this mean square as the denominator. Such a test is called
a pseudo-F or F' test. The real problem here is to determine the degrees
of freedom for the denominator mean square. According to Bennett and
Franklin, if
MS = GI(MS)I + G2(MS)2 + • • •
and (MS)i is based on df, (MS )2 is based on Vj df, and so on, then the
degrees of freedom for MS are
(MS)^
+ ail{MS)l/v2] + ‘ '
(0.0138)^
“ (1)^[(0.0107)V4] + (1)2[(0.0056)V2] + (-1)"[(0.0025)"/4]
_ 1.9044 X 10“^
“ 0.4586 X 10"'‘ ^
MSG 0.7866 _ ^
“ " 0.0138 “ ■
with 2 and 4.2 df, which is significant at the 1 percent level of significance
based on F with 2 and 4 or 2 and 5 df.
10.7 REMARKS
A glance at these EMS values will show that there is no test for the main
effects A. and .6 in a fixed model, as interaction and error are hopelessly con-
founded. The only test possible is to assume that there is no interaction; then
= 0, and the main effects are tested against the error. If a no-interaction
assumption is not reasonable from information outside the experiment, the
investigator should not run one observation per cell but should replicate the
data in a fixed model.
For a random model both main effects can be tested whether interaction
is present or not. For a mixed model there is a test for the fixed effect A but
224 Fixed, random, and mixed models
no test for the random effect B. This may not be a serious drawback, since
the fixed effect is often the most important; the B effect is included chiefly for
reduction of the error term. Such a situation is seen in a randomized block
design where treatments are fixed, but blocks may be chosen at random.
In the discussion of the single-factor experiment it was assumed that
there were equal sample sizes n for each treatment. If this is not the case, it
can be shown that the expected treatment mean square is
or + no(j>.
and
ttn =
S;=i nj
(k - l)N
where
N= Z yi;
J= 1
The test for treatment effect is to compare the treatment mean square with
the error mean square; the use of HQ is primarily for computing components
of variance.
PROBLEMS
10.1 An experiment is run on the effects of three randomly selected operators and five
fixed aluminizers on the aluminum thickness of a TV tube. Two readings are
made for each operator-aluminizer combination. The following ANOVA table
is compiled.
Source df SS MS
Totals 29 563,030
Assuming complete randomization, determine the EMS column for this problem
and make the indicated significance tests.
10.2 Consider a three-factor experiment where factor T is at u levels, factor B at
h levels, and factor C at c levels. The experiment is to be run in a completely
randomized manner with ti observations for each treatment combination. As-
suming factor A is run at a random levels and both B and C at fixed levels, deter-
mine the EMS column and indicate what tests would be made after the analysis.
10.3 Repeat Problem 10.2 with A and B at random levels, but C at fixed levels.
Problems 225
10.4 Repeat Problem 10.2 with all three factors at random levels.
10.5 Consider the completely randomized design of a four-factor experiment similar
to Example 10.2. Assuming factors A and B are at fixed levels and C and D are
at random levels, set up the EMS column and indicate the tests to be made.
10.6 Assuming three factors are at random levels and one is at fixed levels in Prob-
lem 10.5, work out the EMS column and the tests to be run.
10.7 A physical education experiment will be conducted to investigate the effects of
four types of exercise on heart rate. Five subjects are chosen at random from a
physical education class. A subject does each exercise twice. The 40 measurements
are done in a completely randomized order with each subject allowed at least
10 minutes’ rest between exercises.
a. Give the model. State the parameters of its random variables.
b. Work out the expected mean squares of all the effects. Also give the degrees
of freedom for the effects.
c. Tell how to form the F’s for the tests that can be made.
10.8 Four elementary schools are chosen at random in a large city system and two
methods of instruction are tried in the third, fourth, and fifth grades,
a. Outline an ANOVA layout to test the effect of schools, methods, and grades
on gain in reading scores assuming ten students per class.
b. Work out the EMS column and indicate what tests are possible in this
situation.
10.9 As part of an experiment on testing viscosity four operators were chosen at
random and four instruments were randomly selected from many instruments
available. Each operator was to test a sample of product twice on each instru-
ment. Results showed:
Source df MS
0, 3 1498
Ij 3 1816
OIu 9 218
16 67
Determine the EMS column and find the percentage of the total variance in
these readings that is attributable to each term in the model, if it is significant.
10.10 Six different formulas or mixes of concrete are to be purchased from five com-
peting suppliers. Tests of crushing strength are to be made on two blocks con-
structed according to each formula-supplier combination. One block will be
poured, hardened, and tested for each combination before the second block is
formed, thus making two replications of the whole experiment. Assuming mixes
and suppliers fixed and replications random, set up a mathematical model for
this situation, outline its ANOVA table, and show what F tests are appropriate.
10.11 Determine the EMS column for Problem 3.6 and solve for components of
variance.
226 Fixed, random, and mixed models
10.12 Derive the expression for HQ in the EMS column of a single-factor completely
randomized experiment where the ns are unequal.
10.13 Verify the numerical results of Table 10.4 using the methods of Chapter 6.
10.14 Verify the EMS column of Table 10.4 by the method of Section 10.4.
10.15 Set up F' tests for Problem 10.4 and explain how the df would be determined.
10.16 Set up F' tests for Problem 10.5 and explain how the df would be determined.
II
Nested and Nested-Factorial Experiments
11.1 INTRODUCTION
Example 11.1 In a recent in-plant training course the members of the class
were assigned a final problem. Each class member was to go into the plant
and set up an experiment using the techniques that had been discussed in
class. One engineer wanted to study the strain readings of glass cathode
supports from five different machines. Each machine had four “heads” on
which the glass was formed, and she decided to take four samples from each
head. She treated this experiment as a 5 x 4 factorial with four replications
per cell. Complete randomization of the testing for strain readings presented
no problem. Her model was
Her data and analysis appear in Table 11.1. In this model she assumed that
both machines and heads were fixed, and used the 10 percent significance
level. The results indicated no significant machine or head effect on strain
readings, but there was a significant interaction at the 10 percent level of
significance.
The question was raised as to whether the four heads were actually
removed from machine A and mounted on machine B, then on C, and so
on. Of course, the answer was “no” as each machine had its own four heads.
Thus machines and heads did not form a factorial experiment, as the heads
on each machine were unique for that particular machine. In such a case
the experiment is called a nested experiment: levels of one factor are nested
within, or are subsamples of, levels of another factor. Such experiments are
also sometimes called hierarchical experiments.
227
228 Nested and nested-factorial experiments
Machine
Head A B C D E
1 6 10 0 11 1
2 9 0 0 4
0 7 5 6 7
8 12 5 4 9
2 13 2 10 5 6
3 1 11 10 7
9 1 6 8 0
8 10 7 3 3
3 1 4 8 1 3
10 1 5 8 0
0 7 0 9 2
6 9 7 4 2
4 7 0 7 0 3
4 3 2 8 7
7 4 5 6 4
9 1 4 5 0
Totals 79 969.95
Table 11.2 Data for Strain- Reading Pro' :j>lem in a Nested Arrangement
C _. A.. . .
rOA' ^ ' 1
Machine A B c 1 D E
Headci>i'>>J 1 2 3, 4 5 6 1- 8 1
9 10 11 12 13 14 15 16 17 18 19 20
6 13 1 ' 7 10 2 4 0 0 10 8 7 11 5 1 0 1 6 3 3
2 3 10 4 9 1 1
3 ( 0 11 5 2 0 10 8 8 4 7 0 7
0 9 0 7 7 1 7 4 5 6 0 5 6 8 9 6 7 0 2 4
8 8 6 9 12 10 9 1 5 7 7 4 4 3 4 5 9 3 2 0
Head
totals 16 33 17 27 38 14 21 8
1 10 34 20 18 21 26 22 19 21 16 7 14
Machine
totals 93 81 82 88 58
As the heads that are mounted on the machine can be chosen from
many possible heads, we might consider the four heads as a random sample
of heads that might be used on a given machine. If such heads were selected
at random for the machines, the model would be
This nested model has no interaction present, as the heads are not
crossed with the five machines. If heads are considered random and ma-
chines fixed, the proper EMS values can be determined by the rules given
in Chapter 10, as shown in Table 11.3.
This breakdown shows that the head effect is to be tested against the
error, and the machine effect is to be tested against the heads-within-machines
effect.
To analyze the data for a nested design, first determine the total sum
of squares
SS,„,„ = 6^ + 2^ + • • • + 4^ + 0^ - = 969.95
5 4 4
F R R
Source i j k EMS
Mi 0 4 4 + 4cr^ +
Hjw 1 1 4 af + 4a^
^kiU) 1 1 1
230 Nested and nestedrfactorial experiments
Machine A:
(16)2 (33^2 (27)2 (93)2
SS„ = = 50.19
4 16
Machine B :
(38)2 (14)2 (21)2 (g)2 (81)2
SS„ = = 126.18
4 16
Machine C:
(10)2 + (34)2 (2Q)2 (lg)2 (82)2
SS„ : - 74.75
4 16
Machine D :
(21)2 ^ (26)2 + (22)2 ^ (19)2 (88)2
SS„ : = 6.50
4 16
Machine E :
GO
77i
Totals 79 969.95
to the method by which this term was computed, one can look at its parti-
tioning into five parts with 3 df per machine, and since all parts are or-
thogonal, a further breakdown and tests of Table 11.4 can be made as in
Table 11.5.
Source df SS MS F FO.90
From Appendix Table E.l (although one should have a 10 percent table)
p: 2 3 4
2.83 3.40 3.74
LSR: 4.64 5.58 6.13
One should then examine head 5 as its mean is larger than two of the
other three.
For machine C means are 2.50, 4.50, 5.00, 8.50 for heads 9, 12, 11, 10,
and since the largest gap is 8.5 — 2.5 = 6, which is less than 6.13, no signifi-
cant differences are detected. This is so because the differences are significant
at the 10 percent level and not at the 5 percent level on which Table E.l is
based. One might still examine the heads on machine C for lack of consis-
tency. In a nested model it is also seen that the 5 SS between heads within
each machine are “pooled,” or added, to give 282.87. This assumes that these
sums of squares (which are proportional to variances) within each machine
are of about the same magnitude. This might be questioned, as these sums
of squares are 50.19, 126.18, 74.75, 6.50, and 25.25. If these five are really
different, it appears as if the greatest variability is in machine B. This should
indicate the need for work on each machine with an aim toward more
homogeneous strain readings between heads on each machine. This example
shows the importance of recognizing the difference between a nested experi-
ment and a factorial experiment.
Yijk = + ^kiij)
or
Yijk = fi + (/9.. - /^) + - I^L.) + {Yijk -
which is an identity.
Using the best estimates of these population means from the sample
data, the sample model is
i j k
- y.f = I
i= 1
- T,y + ti,'kyu. -
i j
' j k
as the sums of cross products equal zero. This expresses the idea that the
total sum of squares is equal to the sum of squares between levels of A, plus
11.4 Nested-factorial experiments 233
the sum of squares between levels of B within each level of A, plus the sum
of the squares of the errors. The degrees of freedom are
Source df SS MS
■A Tf SS^
Ai a — \
^ nb nab a — 1
a b rj~i 2 fl
TL SSs
^j(t) a{b - 1)
i j i nb a{b - 1)
abn Q b rj-f 2
SS,
P, ^ . .V ab{n — 1) ZSZ
i J k
-
I J ab{n — 1)
abn rp 2
1
Totals abn — 1 EEL
i j k
-
nab
This is essentially the form followed in the problem of the last section.
Note that the
a b
Tfiij- “ Tf
n
yt nb
I J
a / b ryl
" ij- Tl\
" n nb )
which shows the way in which the sum of squares was calculated in the last
section: by getting the sum of squares between levels of B for each level of
A and then pooling over all levels of A.
Yijkm = + Gj + MGij +
where
Mj = methods, i = 1, 2
Gj = groups,; = 1, 2, 3
= teams within groups, k = 1, 2, 3 for all j
£m{ijk) = random error, m = 1, 2 for all k
The EMS values are shown in Table 11.7, which indicates the proper F tests
to run.
2 3 3 2
F F R R
Source i 1 k m EMS
M, 0 3 3 2 + + 18^^
Gj 2 0 3 2 + 4a} -f 12}>Q
MG,j 0 0 3 2 +
T^kU) 2 1 1 2 + 4a}
^Gku) 0 1 1 2 +
^m(ijk) 1 1 1 1
11.4 Nested-factorial experiments 235
Group I II III
Team 1 2 3 4 5 6 7 8 9
Method I 20.2 26.2 23.8 22.0 22.6 22.9 23.1 22.9 21.8
24.1 26.9 24.9 23.5 24.6 25.0 22.9 23.7 23.5
Method II 14.2 18.0 12.5 14.1 14.0 13.7 14.1 12.2 12.7
16.2 19.1 15.4 16.1 18.1 16.0 16.1 13.8 15.1
Source df SS MS EMS
M, 1 651.95 651.95 +
Gj 2 16.05 8.02 + A-(7Y ”1“ \^(J)Q
MG,j 2 1.19 0.60 ol -f
6 39.26 6.54 ol + 4(TJ
6 10.72 1.79 ol +
18 41.59
2
2.31 0-£
Totals 35 760.76
Team
Method
1 2 3 Totals
Team
Method
4 5 6 Totals
Team
Method
7 8 9 Totals
(241.5)'
= 5116.39 - 4860.21 = 256.18
(146.1)2 -f (95.4)'
SS method 4860.21 - 214.19
(74.7)2 + (90.2)2
SS = 4860.21 - 35.74
For Group 11
(45.5)2 + (47.2)2 (479)2 4. + (32.1)^ + (29.7)^
SS cell
2
(232.6)'
= 199.96
12“
(140.6)2 + (92.0)'
SS method - 4508.56 = 196.83
Team 13 2
Mean 18.68 19.15 22.55
Many statisticians who work with psychologists and educators on the design
of their experiments treat repeated-measures designs as a unique topic in
these fields of application of statistics. It is the purpose of this section to
show that these designs are but a special case of factorial and nested-factorial
experiments. Numerical examples and mathematical models are used to
illustrate the correspondence between these designs.
One of the simplest of these cases is when a pretest and posttest are
given to the same group of subjects after a certain time lapse during which
some special instruction may have been administered.
1 100 115
2 110 125
3 90 105
4 no 130
5 125 140
6 130 140
7 105 125
Since the same seven subjects were given each test, we have two repeated
measures on each subject. Winer [24, p. 266] would handle this as a between-
subjects, within-subjects repeated-measures design and report results as
shown in Table 11.14.
240 Nested and nested-factorial experiments
Source
of Variation df SS MS F
Totals 13 2985.71
^ ^ + ^jii)
withdf: 6 7
And then the within-subjects’ data are further broken down into tests and
residual. This residual is actually the subject by test interaction so the
model becomes
Yij = ^ + Tj + STij
with df: 6 16
This model contains no error term since only one pretest score and one post-
test score are available on each subject.
Now if this problem were considered in more general terms instead of
the so-called repeated-measures format, one would consider this data layout
as a two-factor factorial experiment with one observation per treatment.
Subjects should be chosen at random and tests should be fixed. Using the
algorithm given in Chapter 10, the resulting ANOVA layout would be as
shown in Table 11.15.
Here k = 1 and a separate error is not retrievable. A glance at the
EMS column indicates that the only proper F test is to compare the test
mean square with the S x T interaction mean square, which was the test
as shown in Table 11.14.
7 2 1
R F R
Source df i i k EMS
5, 6 1 2 1 ol + IGI
TJ 1 7 0 1 T ^ST T
S7-,, 6 1 0 1 ol + ojj
0 1 1 1 (Tg (not retrievable)
11.6 Repeated-measures design and nested-factorial experiments 241
= (12.05)^ = 145 = F
Example 11.4 A slightly more complex problem involves three factors
where the subjects are nested within groups. These groups could be classes
or experimental conditions and then again each subject is subjected to
repeated measures. The study cited on physical strength was extended so
that three groups of subjects were involved, two being subjected to special
experimental training and the third acting as a control with no special
training. Again each subject was given a pre- and posttest where the
measurement here was velocity of a baseball throw in meters per second.
The resulting data are given in Table 11.16. If these data are treated as a
1 26.25 29.50
2 24.33 27.62
3 22.52 25.71
I
4 29.33 31.55
5 28.90 31.35
6 25.13 29.07
7 29.33 31.15
8 27.47 28.74
9 25.19 26.11
10 23.53 25.45
II 11 24.57 25.58
12 26.88 27.70
13 27.86 28.82
14 28.09 28.99
15 22.27 22.52
16 21.55 21.79
17 23.31 23.53
18 30.03 30.21
III
19 28.17 28.65
20 28.09 28.33
21 27.55 27.86
242 Nested and nested-factorial experiments
Source df SS MS EMS F
F R F R
3 1 2 1
Source df i i k m EMS
G, 2 0 1 2 1 G^ + 2(7| + 140G
^j(i} 18 1 1 2 1 GI + 2cr|
u 1 3 7 0 1 + ^TS + 210^
GT,, 2 0 7 0 1 Gg + (7^5 + l<pQj
18 1 1 0 1 G^ + Gjg
0 1 1 1 1 Gg (not retrievable)
This scheme generates the EMS column that was simply reported in
Table 11.17, and that column indicates the proper tests to be made. The
numerical results are, of course, the same.
The nested-factorial experiment covers many situations in addition to
the repeated-measures design. The factor in the nest may be farms within
11.6 Repeated-measures design and nested-factorial experiments 243
Case I:
between
subjects
within subjects
. ^ s
+ Bk-h ABik + BSkji^i^-\- C^-\- ACim + CSmkjii) + + ABC ikm T BCSj^kj{i)
Case II:
between subjects
, ^,
Yijkm = + AB^j + Sk(ij)
within subjects
No error terms have been included since they are not retrievable.
244 Nested and nested-factorial experiments
11.7 SUMMARY
The summary at the end of Chapter 6 may now be extended for Part 11.
PROBLEMS
11.1 Porosity readings on condenser paper were recorded for paper from four rolls
taken at random from each of three lots. The results were as follows. Analyze
these data, assuming lots are hxed and rolls random.
Problems 245
Lot I II III
Roll 1 2 3 4 5 6 7 8 9 10 11 12
1.5 1.5 2.7 3.0 1.9 2.3 1.8 1.9 2.5 3.2 1.4 7.8
1.7 1.6 1.9 2.4 1.5 2.4 2.9 3.5 2.9 5.5 1.5 5.2
1.6 1.7 2.0 2.6 2.1 2.4 4.7 2.8 3.3 7.1 3.4 5.0
11.2 In Problem 11.1 how would the results change (if they do) if the lots were chosen
at random?
11.3 Set up the EMS column and indicate the proper tests to make if /I is a fixed
factor at hve levels, B is nested within A at four random levels for each level
of A, C is nested within B at three random levels, and two observations are
made in each “cell.”
11.4 Repeat Problem 11.3 for A and B crossed or factorial and C nested within the
A, B cells.
11.5 Two types of machines are used to wind coils. One type is hand operated; two
machines of this type are available. The other type is power operated; two
machines of this type are available. Three coils are wound on each machine
from two different wiring stocks. Each coil is then measured for the outside
diameter of the wire at a middle position on the coil. The results were as follows.
(Units are 10“inch.)
Machine Number 2 3 5 8
Set up the model for this problem and determine what tests can be run.
11.6 Do a complete analysis of Problem 11.5.
11.7 If the outside diameter readings in Problem 11.5 were taken at three fixed posi-
tions on the coil (tip, middle, and end), what model would now be appropriate
and what tests could be run?
11.8 An experiment is to be designed to compare faculty morale in three types of
junior high school organizations: grades 6-8, grades 7-9, and grades 7-8. Two
schools are randomly selected to represent each of these three types of organiza-
tions and data are to be collected from several teachers in each school. Some-
one suggests that the morale of male and female teachers might differ so it is
246 Nested and nested-factorial experiments
agreed to choose random samples of five male and five female teachers from each
school so that sex differences in morale may be checked out too.
a. Set up a mathematical model for this experiment, outline into ANOVA, and
indicate what tests can be made.
b. Make any recommendations that seem reasonable based on your answers to
part (a), assuming that the data have not been collected as yet.
11.9 In an attempt to study the effectiveness of two corn varieties {V^ and V2), two
corn-producing counties (C) in Iowa were chosen at random. Then four farms
(F) were randomly selected within each county. Seed from both varieties were
sent to these farms and planted in random plots. When harvested, the number
of bushels of corn per acre was recorded for each variety on each farm.
a. Write a mathematical model for this situation.
b. Determine the EMS for varieties in this experiment.
c. Assuming only counties (C) showed significantly different average yields, and
MSc =130 and MS^ = 10 (farms or error), find the percentage of the vari-
ance in this experiment that can be attributed to county differences.
11.10 An educator proposes a new teaching method and wishes to compare the achieve-
ment of students using his method with that of students using a traditional
method. Twenty students are randomly placed into two groups with ten students
per group. Tests are given to all 20 students at the beginning of a semester, at
the end of the semester, and ten weeks after the end of the semester. The educator
wishes to see whether there is a difference in the average achievement between
the two methods at each of the three time periods.
a. Write a mathematical model for this situation.
b. Set up an ANOVA table and show the F tests that can be made.
c. If the educator’s method is really better than the traditional method, what
would you expect if you graphed the results at each time period? Show by
sketch.
d. If method A—the new method—gave a mean achievement score across all
time periods of 60.0, explain in some detail how you would set confidence
limits on this method mean.
11.11 In Example 11.4 in the text the ANOVA of Table 11.17 suggests that further
analysis is needed. Using the data of Table 11.16, make further analyses and
state your conclusions. (A graph might also be helpful.)
11.12 In studying the intellectual attitude toward science as measured on the Bratt
attitude scale, a science educator gave a pretest and posttest to an experimental
group exposed to a new curriculum and a control group that was not so exposed.
There were 15 students in each group. Set up a mathematical model for this
situation and determine what tests are appropriate.
11.13 In Problem 11.12 a strong interaction was found between the groups and the
tests with means as follows: pretest control = 57.00, pretest experimental =
54.96, posttest control = 54.66, posttest experimental = 65.66. Assuming the
sum of squares for the error term used to test for this interaction was 492.61,
make further tests and state some conclusions based on this information.
Problems 247
11.14 A researcher wished to test the figural fluency of three groups of Egyptian stu-
dents: those taught by the Purdue Creative Thinking Program (PCTP) in a
restricted atmosphere, those taught by PCTP in a permissive atmosphere, and
a control group taught in a traditional manner with no creativity training.
Four classes were randomly assigned to each of the three methods. Before
training began, all pupils were administered a Group Embedded Figures Test
and classified as either field-dependent of field-independent students. After six
weeks of training results on the averages of students within these classes on
this one subtest (figural fluency) were found to be
Source df SS MS F Prob.
Source df SS MS F Prob.
Discuss these results and explain how the F tests were made.
11.16 For Problem 11.15 it was known in advance that the experimenter wished to
compare the two experimental curricula results with the control, and also to
compare the two experimental curricula results with each other. Set up proper
a priori tests for this purpose and test the results based on the following means
and the table given in Problem 11.15.
248 Nested and nested-factorial experiments
Day 1 2 3
Grader A B A B A B
Sample 1 4 11 5 11 0 6
2 6 7 17 13 -1 -2
3 6 10 8 15 2 5
4 13 11 3 14 8 2
5 7 10 14 20 8 6
6 7 11 11 19 4 10
7 14 16 6 11 5 18
8 12 10 11 17 10 13
9 9 12 16 4 16 17
10 6 9 -1 9 8 15
11 8 13 3 14 7 11
Assuming graders fixed, days random, and samples within days random, set up
a mathematical model for this experiment and determine the EMS column.
11.20 Complete the ANOVA for Problem 11.19 and comment on the results.
11.21 In a filling process two random runs are made. For each run six hoppers are
used, but they may not be the same hoppers on the second run. Assume that
hoppers are a random sample of possible hoppers. Data are taken from the
bottom, middle, and top of each hopper in order to check on a possible position
effect on filling. Two observations are made within positions and runs on all
hoppers. Set up a mathematical model for this problem and the associated
EMS column.
11.22 The data below are from the experiment described in Problem 11.21.
Problems 249
Run 1 Run 2
1 19 0 31 25 25 13 6 6 0 6 0 0
2 13 0 19 38 31 25 0 13 0 31 6 0
3 13 0 31 19 13 0 28 6 25 19 13 28
4 13 0 19 31 25 13 28 0 6 0 28 25
5 19 19 19 38 19 25 16 13 13 19 0 0
6 6 0 0 0 0 0 13 6 6 19 0 19
10 p p 00 't; P 't
420.2
00
• HH ■>—1 P P —H m P
• l—H r4 nl P P
rn r^i
p p p p p p 1
m 00 p
r^i
HH od r-' od n 't' P p P r~i P o
D in 'it •
a
HH o p p p T—( p 00 p p OO oo
o HH cK P P P 0 P 0 p P P o\
m rn (N
tri p p 00 p p p P ON p
210.7
HH in in NO NO P P P P p o^
-.|<N (N
> oo p ON P P in p 1
p OO <N r^i
od od NO 'It P NO P 0 P P P o
(N <N fN
HH '•O Cs| (N p p 0 m p NO p O)
HH Md MD NO P P P P P P P P r-'
HH
(N
1
p p P p P p 'it p 'll; p p p
156.2
1
i '' G in in 'ij’ 't '=t P 'It P '=t 't P m’
1
o^ in in OO p 00 p NO 'It
CNI
HH 'sd NO 0 NO P 't 06 P P
CNI
c3
•
c:
B h-^ p P 00 p p 00 Csl in
3 HH in P p 't' P P P P P P od
<
P p p ■It m p
163.3
T-H 00 m
G in in in n n P 't P \6 P P
p p NO ON 00 ON <N ON m p p IT)
H—1
r-' NO 0 NO P P P P P P o
nj <N
c
_o '-H CNI m
Totals
+-> (U 0 0
UH "HH ’p ■DH
+-> c c c c
c c c c
a> d d
o '4—> 0:1 00 00
C c/5
o o
U OH
250
h—(
t ■ p lO p p p 1 oo p p p p
h-H r4 r4 (N 'O I ( T—H T—H r4 rn r4 od
h-H lO p p p p p lO p p p p oo
h-^ T—H 1—H cn T—H ^H 1 ( <ni
lO
oq p p m p p p p p p m
h“H
r4 <N 1 (
Copper
C (N (N T—H MD <N (N (N od
HH oo OO p o p m 1 ^ p T—H p m
HH O rn m' o (N rn (N OO rn rn T—H
HH T-H
- (N oo p p p p p p
HH 1
HH r4 O (N 'vO <N c4 \6 rn
rq p p p p p P p p p
1 ^
(N (N m’ oo’ (N rd r4 rn rn rn cK
l-H p p p p p
1
p p T—H T—H
HH m p p p P OO p 1 OO p » ^ p p p
h-H rn O r4 T—H T—H rn
Iron Oxide
HH
oq p p p p p p p
1 ^•
r ■ "H r4 r4 (N (N 'sO T—H T—H
HH (N 1—! rq 'O p P p p p p o^
HH rn rn O r4 m' H H (N (N Md
HH
r'« H
HH oc P p p p 1 p p p p p
-H|(N HH
r4 ini mi
'O
p oc p OO p p p ^H o^ p 'O
m' m’ rc rn ro <N oo (N fd T—H Md
h-H p p p (N p p 1 ^
p p p (N
HH
HH r4 (N <N r^’ 1p rd oi 'O
HH ■vO P p p T—H f ^ o^ 1 ^ p r- p p
H o m' o rn
mi
-
Iron Filings
p oq p m p p m ■^H p
1 ^
HH m’ rn (N O (N (N ' ' 'O r4 <N r4 r-’
T-H
HH m p p r-; p p p p P p p
< !
HH 1 (N r4 lo r4 mi
HH o\ P oo oo oo in T—H p
-i|(N HH
1—H O m' o (N o o o fN T—H
P p p p m 1 oo p p p p
r4 r4 1—H T—H ^H P rn r4 H od
Concentration
Sample 2
Sample 3
Sample 1
Totals
Position
251
12
Experiments of Two or More Factors —
Restrictions on Randomization
12.1 INTRODUCTION
In the discussion of factorial and nested experiments it was assumed that
the whole experiment was performed in a completely randomized manner.
In practice, however, it may not be feasible to run the entire experiment
several times in one day, or by one experimenter, and so forth. It then becomes
necessary to restrict this complete randomization and block the experiment
in the same manner that a single-factor experiment was blocked in Chapter 4.
Instead of running several replications of the experiment all at one time,
it may be possible to run one complete replication on one day, a second
complete replication on another day, a third replication on a third day, and
so on. In this case each replication is a block, and the design is a randomized
block design with a complete factorial or nested experiment randomized
within each block.
Occasionally a second restriction on randomization is necessary, in
which case a Latin square design might be used, with the treatments in the
square representing a complete factorial or nested experiment.
252
12.2 Factorial experiment in a randomized block design 253
^ijkm 1
““ Rfi '^ij 4 “ £,n(ijk)
where Rj^ represents the blocks or replications and ijj represents the treat-
ments. If the treatments are formed from a two-factor factorial experiment,
then
^ij =
Source df
Replications 3
Treatments 5
Ai 2
1
AB,^ 2
Error eiji, 15
Total 23
254 Experiments of two or more factors—restrictions on randomization
in Table 12.2. Here the factors are fixed and replications are considered as
random.
3 2 4 1
F F R R
Source df i j k m EMS
3 3 2 1 1 + 6o-|
At 2 0 2 4 1 + 2(T|^ + A
RA,, 6 0 2 1 1 + 2cr|^
Bj 1 3 0 4 1 G^ + 3cr^g + 12 (t)B
RBjk 3 3 0 1 1 GI + 2G\B
AB,j 2 0 0 4 1 + ^RAB + ^(I^AB
RABijk 6 0 0 1 1 + G^^B
^m(ijk) 0 1 1 1 1 G^ (not retrievable)
Total 23
Since the degrees of freedom are quite low for the tests indicated above,
and since there is no separate estimate of error variance, it is customary to
assume that = (^RAB = 0 to pool these three terms to serve
as the error variance. The reduced model is then as shown in Table 12.3.
Here all effects are tested against the 15-df error term. Although we usually
pool the interactions of replications with treatment effects for the error term,
this is not always necessary. It will depend upon the degrees of freedom
available for testing various hypotheses in Table 12.2, and also whether or
not any repeat measurements can be taken within a replication. If the latter
is possible, a separate estimate of aj is available and all tests in Table 12.2
may be made.
Source df EMS
Rk 3 erf + 6o-|
At 2 + 8^^
Bj 1 + 12^5
AB,j 2 + ^^AB
^ijk 15
Total 23
12.2 Factorial experiment in a randomized block design 255
Operator 1 2 3 4 5
Nozzle A B C A B C A B C A B C A B C
Replication
I 6 13 10 26 4 -35 11 17 11 21 -5 12 25 15 -4
II 6 6 10 12 4 0 4 10 -10 14 2 -2 18 8 10
The levels are random, Oj levels are random, and levels are fixed. If m
repeat measurements are considered in the model, but m = i, the EMS
values are those appearing in Table 12.5.
3 5 3 1
F R R R
Source df i j k m EMS
Rk 2 3 5 1 1 15(T|
Ni 2 0 5 3 1 G^ + 3crj5o + 15^;^^
Oj 4 3 1 3 1 G^, + 9Gi
NO,j 8 0 1 3 1 G^ + 2G^O
^majk) 28 1 1 1 1 <^1
Total 44
256 Experiments of two or more factors—restrictions on randomization
These EMS values indicate that all main effects and the interaction can
be tested with reasonable precision (df). The sums of squares are computed
in the usual manner from the subtotals in Table 12.6.
Operator
Nozzle
Replication Total Nozzle 1 2 3 4 5 Total
I 127 A -3 43 19 42 68 169
II 92 B 32 19 44 -8 24 111
III 29 C 9 -49 -16 -6 30 -32
Using these subtotals, the usual sums of squares can easily be computed
and the results displayed as in Table 12.7.
Totals 44 7085.24
MS = MS^K + - MS^^o
12.2 Factorial experiment in a randomized block design 257
nozzle A
Cell
Totals
Source df SS MS EMS F
Totals 44 7085.24
Operator Nozzle
Here the coefficients are 1,1, and — 1, and the corresponding mean squares
and degrees of freedom are
713.48
68.06 + 227.68 - 108.21
713.48
3.80
187.53
and
^ (187.53)^
" (l)^[(68.06)V4] + (1)^[(227.68)V8] + (-1)^[(108.21)V16]
With 2 and 4 df the 5 percent F value is 6.94, and with 2 and 5 df it is 5.79;
hence the F' = 3.80 is not significant at the 5 percent level. Hence all F
tests indicate no significant effects. The N x O interaction does not show
significance as it did in Table 12.7, because the test is less sensitive due to the
reduction in degrees of freedom from 28 to 16 in the error mean square.
Since both the N x R and R x O interaction F tests are less than one, we
are probably justified in assuming that and aj^o ^re zero and in pooling
these with for a 28-df error mean square as in Table 12.7.
^ijk = F F Ri Tj Jk F Sijj.
where Xj represents the treatments, and both R^ and y^. represent restrictions
on the randomization. If a 4 x 4 Latin square is used for a given experiment,
such as the one in Section 4.5, the four treatments A, B, C, and D could repre-
sent the four treatment combinations of a 2 x 2 factorial (Table 12.9). Here
L ^ F Bq F ABf^^q
Car
Position I II III IV
1 C D A B
2 B C D A
3 A B C D
4 D A B C
12.4 Remarks 259
B ). The types might be black- and white-walled tires. The complete model
2
would then be
Source df
R. 3
yk 3
Am i]1
Rq 1 y 3 df for T :
ij1
c
^ikmq 6
Total 15
12.4 REMARKS
and when this experiment has one restriction on the randomization, the
model is
^ijk = + Fj +
^ -j
V
where the whole nested factorial is run in a randomized block design. The
analysis of such designs follows from the methods of Chapter 11.
260 Experiments of two or more factors—restrictions on randomization
12.5 SUMMARY
3. Latin square
a. Complete a. Factorial ANOVA with
Yijkm = B + + replications and
+ Aj- + Bj + AB^j positions
d" jkm
B. Nested
(hierarchical)
1. Completely randomized 1. Nested ANOVA
^ijk = B + A,-
+ Bjii) + h{ij)
2. Randomized block 2. Nested ANOVA with
blocks R,
a. Complete a. Nested ANOVA with
TjVc = + ^/c + A,- blocks R,
+ + ^ijk
3. Latin square 3.
a. Complete a. Nested ANOVA with
Tj/cm ^ B + J^k + Jm blocks and positions
+ A,- + Bj^i^ +
C. Nested
factorial
1. Completely randomized 1. Nested-factorial ANOVA
bjfcm = /^ + A,■ +
+ C, + AC,,
+ ^^kj(i) + ^m{ijk)
Problems 261
2. Randomized block 2.
a. Complete a. ANOVA with blocks
j{i) T £,•
3. Latin square 3.
a. Complete a. ANOVA with blocks
and positions
+ + BCqji^i^
T £(' jkmq
PROBLEMS
12.1 Data on the glass rating of tubes taken from two fixed stations and three shifts are
reeorded each week for three weeks. Considering the weeks as blocks, the six
treatment combinations (two stations by three shifts) were tested in a random
order each week, with results as follows:
Station 1 Station 2
Shift 1 2 3 1 2 3
Week 1 3 3 3 6 3 6
6 4 6 8 9 8
6 7 7 11 11 13
Week 2 14 8 11 4 15 4
16 8 12 6 15 7
19 9 17 7 17 10
Week 3 2 2 2 2 2 10
3 3 4 5 4 12
6 4 6 7 6 13
representing the four combinations of two ply values and two tread types. Analyze
the data for this revised design—a 2^ factorial run in a Latin square design.
262 Experiments of two or more factors—restrictions on randomization
12.3 If in Problem 11.5 the stock factor is replaced by “days,” considered random,
and the rest of the experiment is performed in a random manner on each of the
two days, use the same numerical results and reanalyze with this restriction.
12.4 Again, in Problem 11.5 assume that the whole experiment (24 readings) was
repeated on five randomly selected days. Set up an outline of this experiment,
including the EMS column, df, and so on.
12.5 Three complete replications were run in a study of aluminum thickness. Samples
were taken from five machines and two types of slug—rivet and staple—in each
replicate. Considering replicates random and machines and type of slug fixed,
set up a model for this problem and determine what tests can be made. Comment
on the adequacy of the various tests. Data are as follows:
Machine
Replication Slug 1 2 3 4 5
12.6 Analyze the data of Problem 12.5 using a computer program and state your
conclusions.
12.7 Data were collected for three consecutive years on the number of days after
planting that the first flowering of a plant occurred. Each year five varieties of
plants were planted at six different stations. Considering the repeat years random
and varieties and stations fixed, outline the ANOVA for this problem.
12.8 The data for Problem 12.7 are given below.
Variety
Broadchalke
1932 57 67 95 102 123
1933 46 72 90 88 101
1934 28 66 89 109 113
Total 131 205 274 299 337
Problems 263
Variety
Bratton
1932 26 44 92 96 93
1933 38 68 89 89 no
1934 20 64 106 106 115
Total 84 176 287 291 318
Lenham
1932 48 61 78 99 113
1933 35 60 89 87 109
1934 48 75 95 113 111
Total 131 196 262 299 333
Dorstone
1932 50 68 85 117 124
1933 37 65 74 93 102
1934 19 61 80 107 118
Total 106 194 239 317 344
Coaley
1932 23 74 105 103 120
1933 36 47 85 90 101
1934 18 69 85 105 111
Total 77 190 275 298 332
Ipswich
1932 39 57 91 102 112
1933 39 61 82 93 104
1934 43 61 98 98 112
Total 121 179 271 293 328
12.10 Data for Problem 12.9 are given below for two runs, three positions {B, M, T),
six heads (H), and two cycles (C) with each combination:
Run 1 Run 2
B M T B M T
H c, C2 Cl C2 Cl C2 Cl C2 Cl C2 Cl C2
1 19 0 31 25 25 13 6 6 0 6 0 0
2 13 0 19 38 31 25 0 13 0 31 6 0
3 13 0 31 19 13 0 28 6 25 19 13 28
4 13 0 19 31 25 13 28 0 6 0 28 25
5 19 19 19 38 19 25 16 13 13 19 0 0
6 6 0 0 0 0 0 13 6 6 19 0 19
Do a complete analysis of these data. Explain whether or not you could combine
terms in the table as done in the text example.
12.11 Discuss the similarities and differences between a nested experiment with two
factors and a randomized block design with one factor.
12.12 Assuming the nested experiment involving machines and heads in Section 11.2
was completely repeated on three subsequent randomly selected days, outline its
mathematical model and ANOVA table complete with the EMS column.
12.13 For Problem 12.12 show a complete model with all interactions and the reduced
model. Make suggestions regarding how this experiment might be made less
costly and still give about the same information.
12.14 For the nested-factorial example in Section 11.4 on gun loading set up the model
assuming the whole experiment is repeated on four more randomly selected days.
12.15 Examine both the full and reduced models for Problem 12.14 and comment.
13
Factorial Experiment —
Split-Plot Design
13.1 INTRODUCTION
In many experiments where a factorial arrangement is desired, it may not be
possible to randomize completely the order of experimentation. In the last
chapter restrictions on randomization were considered, and both random-
ized block and Latin square designs were discussed. There are still many
practical situations in which it is not at all feasible to even randomize within
a block. Under certain conditions these restrictions will lead to a split-plot
design. An example will show why such designs are quite common.
Example 13.1 The data in Table 13.1 were compiled on the effect of oven
temperature T and baking time B on the life Y of an electrical component
[25].
Looking only at this table of data, we might think of this as a 4 x 3
factorial with three replications per cell and proceed to the analysis in
Baking
Oven Temperature T (°F)
Time B
(min) 580 600 620 640
265
266 Factorial experiment—split-plot design
Source df SS MS EMS
Totals 35 29,331
Table 13.2. Here only the temperature has a significant effect on the life of
the component at the 1 percent significance level.
We might go on now and seek linear, quadratic, and cubic effects of
temperature; but before proceeding, a few questions on design should be
raised. What was the order of experimentation? The data analysis above
assumes a completely randomized design. This means that one of the four
temperatures was chosen at random and the oven was heated to this tem-
perature, then a baking time was chosen at random and an electrical com-
ponent was inserted in the oven and baked for the time selected. After this
run, the whole procedure was repeated until the data were compiled. Now,
was the experiment really conducted in this manner? Of course the answer is
“no.” Once an oven is heated to temperature, all nine components are in-
serted; three are baked for 5 minutes, three for 10 minutes, and three for 15
minutes. We would argue that this is the only practical way to run the ex-
periment. Complete randomization is too impractical as well as too expen-
sive. Fortunately, this restriction on the complete randomization can be
handled in a design called a split-plot design.
The four temperature levels are called plots. They could be called blocks,
but in the previous chapter blocks and replications were used for a complete
rerun of the whole experiment. (The word “plots” has been inherited from
agricultural applications.) In such a setup temperature—a main effect—is
confounded with these plots. If conditions change from one oven temperature
to another, these changes will show up as temperature differences. Thus, in
such a design, a main effect is confounded with plots. This is necessary be-
cause it is often the most practical way to order the experiment. Now once a
temperature has been set up by choosing one of the four temperatures at
random, three components can be placed in the oven and one component
can be baked for 5 minutes, another for 10 minutes, and the third one for
15 minutes. The specific component that is to be baked for 5, 10, and 15
minutes is again decided by a random selection. These three baking-time
levels may be thought of as a splitting of the plot into three parts, one part
for each baking time. This defines the three parts of a main plot that are called
13.1 Introduction 267
split-plots. Note here that only three components were placed in the oven,
not nine. The temperature is then changed to another level and, three more
components are placed in the oven for 5, 10, and 15 minutes baking time.
This same procedure is followed for all four temperatures; after this the
whole experiment may be replicated. These replications may be run several
days after the initial experiment; in fact, it is often desirable to collect data
from two or three replications and then decide whether more replications
are necessary.
The experiment conducted above was actually run as a split-plot experi-
ment and laid out as shown in Table 13.3.
Baking
Oven Temperature T (°F)
Time B
Replication R (min) 580 600 620 640
called split-plots. Since one main effect is confounded with plots and the other
main effect is not, it is usually desirable to place in the split the main effect
we are most concerned about testing, as this factor is not confounded.
We might think that factor B (baking time) is nested with the main plots
but this is not the case, since the same levels of B are used in all plots. A
model for this experiment would be
The first three variable terms in this model represent the whole plot, and
the RT interaction is often referred to as the whole-plot error. The usual
assumption is that this interaction does not exist, that this term is really an
estimate of the error within the main plot. The last four terms represent the
split-plot, and the RTB interaction is referred to as the split-plot error. Some-
times the RB term is also considered nonexistent and is combined with RTB
as an error term. A separate error term might be obtained if it were feasible
to repeat some observations within the split-plot. The proper EMS values can
be found by considering m repeat measurements where m = 1 (Table 13.4).
3 4 3 1
R F F R
Source df i j k m EMS
Total 35
Note in Table 13.4 that the degrees of freedom in the whole plot add
to 11 as there are 12 whole plots (3 replications x 4 temperatures) and the
degrees of freedom in the split-plot add to 24 as there are 2 df within each
whole plot and hence 2x12 within all whole plots.
Since the error mean square cannot be isolated in this experiment,
+ ^RTB is taken as the split-plot error, and is taken as the
13.1 Introduction 269
Source df SS MS EMS
Totals 35 29,331
4165
14.1 (significant at the 1 percent level)
283
< 1 (not significant)
ET^
434
F 6,12 1.79 (not significant at the 5 percent level)
243
No exact tests are available for testing the replication effect or replica-
tion interaction with other factors, but these effects are present only to reduce
the experimental error in this split-plot design.
The results of this split-plot analysis are not too different from the results
using the incorrect method of Table 13.2, but this split-plot design shows the
need for careful consideration of the method of randomization before starting
270 Factorial experiment—split-plot design
Oven Temperature
Baking
oo
o
5 3 4
10 1 5
15 6 2
The numbers 1, 2,..., 6 indicate the order of the experiment: first set tem-
perature at 600, bake for 10 minutes; then set temperature at 620, bake for
15 minutes; and so on. One notes that the order is scattered throughout
the 12 treatment combinations. In the case of the split-plot design a tempera-
ture is randomly chosen (roll the red die) and then the baking times are
chosen at random within that temperature (roll the green die). One such
randomization is shown for the first six combinations as follows;
Oven Temperature
oo
oo
(-/I
5 3 4
10 1 6
15 2 5
In this latter case it should seem more obvious that temperature is con-
founded with the order in which the experiment is run, as three treatments
are run at 580° first, then three are run at 620° second, and so on. It is hoped
that by replicating the whole experiment several times any effect of order
confounded with temperature would average out. Nevertheless, in a split-
plot design a main effect is confounded.
Some statisticians recommend introducing a restriction error into a
model wherever such restrictions occur; see Anderson and McLean [1].
No such term has been introduced here, and so it is strongly recommended
13.1 Introduction 271
that F tests be run only within the whole plot or within the split-plot and
that mean squares in the whole plot not be compared with mean squares in
the split-plot regardless of the EMS column.
Since this type of design is encountered often in industrial experiments,
another example will be considered.
Temperature T (°C)
-55 25 55
Humidity H (percent)
Strip S Box 50 70 90 50 70 90 50 70 90
Replication 11 1
2
R 2 3
4
3 5
6
4 7
8
5 9
10
The model for this design and its associated EMS relations are set up
in Table 13.7.
From the EMS column tests can be made on the effects of replications,
replications by temperature interaction, replications by humidity interac-
tion, replications by temperature by humidity interaction, strips, and strips
by all other factor interactions. Unfortunately no exact tests are available
for the factors of chief importance: temperature, humidity, and temperature
by humidity interaction. Of course, we could first test the hypotheses that
can be tested, and if some of these are not significant at a reasonably high
level (say 25 percent), assume they are nonexistent and then remove these
terms from the EMS column. For example, if the TS interaction can be
assumed as zero (cr< 0), the temperature mean square can be tested against
the RT interaction mean square with 2 and 6 df. If this is not a reasonable
assumption, a pseudo-F {F') test can be used as discussed in Chapter 10.
Unfortunately, data were available only for the first replication of this
experiment, so its complete analysis cannot be given. The method of analysis
is the same as given in Chapter 5, even though this experiment is rather
complicated. It is presented here only to show another actual example of a
split-plot design.
13.2 A split-split-plot design 273
4 3 3 5 2
R F F R R
Source df i j k m Q EMS
Ei 3 1 3 3 5 2 + 18cr^5 + 90ai
Tj 2 4 0 3 5 2 + 6(7^Ts + 24(7 + 30a^Y ■{“ 120^7’
Whole REj 6 1 0 3 5 2 + ^^RTS + 30(7^7’
plot 2 4 3 0 5 2 + ^^RHs + 24cr^5 + 30a^fj -f- 120 ^77
RH,, 6 1 3 0 5 2 + ^^RHS +
THj, 4 4 0 0 5 2 + ^^RTHS + ^^THS + Idcr^JH + 40
RTH,j, 12 1 0 0 5 2 + ^^RTHS + ^^^RTH
Total 359
Temperature (°C)
145 155 165
1 3 4 (a)
2 8 1 5 7 9 Completely randomized
3 2 6
1 (b)
2 4 6 1 7 9 3 2 8 5 Split-plot
3
1 (c)
2 8 9 7 3 1 2 4 6 5 Split-split-plot
3
Temperature (°C)
145° 155 165
III 1 18.7 16.5 21.8 9.5 8.9 11.5 5.7 4.3 5.8
274
13.2 A split-split-plot design 275
4 3 3 3 1
R F F F R
Source df i j k m q EMS
Lj 2 4 0 3 3 1 + 36
RL,j 6 1 0 3 3 1 +
Split-split-plot 2 4 3 3 0 1 + ^^RM + 36
RM,„ 6 1 3 3 0 1 ^^RM
4 4 0 3 0 1 + ^^RLM +
FLMij„, 12 1 0 3 0 1 + 2(^RLM
4 4 3 0 0 1 + ^^RTM + \2(f)jM
FTMii^jn 12 1 3 0 0 1 + ^^RTM
LTMjj^fn 8 4 0 0 0 1 ^RLTM + ^</>LTM
„2
RLTMijf^^ 24 1 0 0 0 1 -h ^RLTM
^qiijkm) 0 1 1 1 1 1 (not retrievable)
Total 107
The EMS column indicates that F tests can be made on all three fixed
effects and their interactions without a separate error term. Since the effect
of replication is often considered of no great interest and the interaction of
replication with the other factors is often assumed to be nonexistent, this
appears to be a feasible experiment. If examination of the EMS column
should indicate that some F tests have too few degrees of freedom in their
276 Factorial experiment—split-plot design
denominator (some statisticians say less than 6), the experimenter might
consider increasing the number of replications to increase the precision of
the test. In the above experiment, if five replications were used, the whole
plot error RL would have 8 df instead of 6, the split-plot error RLT would
have 16 instead of 12, and so on.
Another technique that is sometimes used to increase the degrees of
freedom is to pool all interactions with replications into the error term for
that section of the table. Here one might pool RT and RLT and RM, RLM
and RTM with RLTM if additional degrees of freedom are believed neces-
sary. One way to handle experiments with several replications is to stop
after two or three replications, compute the results, and see if significance
has been achieved or that the F’s are large even if not significant. Then add
another replication, and so on, increasing the precision (and the cost) of the
experiment in the hope of detecting significant effects if they are there.
The numerical results of the rubber cure rate index are given in Table
13.11.
Source df SS MS
R 3 9.41 3.14
L 2 40.66 20.33*
RL 6 16.11 2.68
T 2 3119.51 1559.76***
RT 6 2.07 0.34
LT 4 4.94 1.24
RLT 12 7.81 0.65
M 2 145.71 72.86***
RM 6 3.29 0.55
LM 4 0.35 0.09
RLM 12 2.93 0.24
TM 4 43.69 10.92***
RTM 12 2.22 0.18
LTM 8 1.07 0.14
RLTM 24 4.97 0.21
13.3 SUMMARY
whole plot
+ Bk A- RBn^ + ABj,^ + RTBjjfc
split-plot
PROBLEMS
13.1 As part of the experiment discussed in Example 13.2 a chamber was set at 50
percent relative humidity and two boxes from each of three strips were inserted
and the pulloff force determined. This was repeated for 70 percent and 90 percent,
the order of humidities being randomized. Two replications of the whole experi-
ment were made and the results follow.
Humidity
Set up a split-plot model for this experiment and outline the ANOVA table,
including EMS.
13.2 Analyze Problem 13.1.
13.3 The following data layout was proposed to handle a factorial experiment de-
signed to determine the effects of orifice size and flow rate on environmental
chamber pressure in millimeters of mercury.
3.5
4.0
4.5
tube. The order of measuring by the four operators is randomized each day for
five days of the week. The results are
Set up the model and outline the ANOVA table for this problem.
13.8 Considering days as fixed, operators as random, and replications as random,
analyze Problem 13.7.
13.9 Explain how a split-plot design differs from a nested design, since both have a
factor within levels of another factor.
13.10 From the data of Table 13.9 on the rubber cure rate index experiment verify
the results of Table 13.11.
13.11 Since two qualitative factors in Problem 13.10 show significance, compare their
means by a Newman-Keuls test.
13.12 Since temperature in Problem 13.10 is a highly significant quantitative effect,
fit a proper polynomial (or polynomials) to the data.
13.13 Redo Table 13.10, assuming that the three laboratories are chosen at random
from a large number of possible laboratories.
13.14 Analyze the results of Table 13.9, based on the EMS values of Problem 13.13.
13.15 In a study of the effect of baking temperature and recipe (or mix) on the quality
of cake, three mixes were used; A, B, and C. One of these three was chosen at
random and the mix made up and five cakes were poured, each placed in a differ-
ent oven and baked at temperatures of 300°, 325°, 350°, 375°, and 400°. After
these five cakes had been baked, removed, and their quality (T) evaluated, an-
other mix was chosen and again five cakes were poured and baked at the five
temperatures. Then this was repeated on the third mix. This experiment was
replicated r times.
a. Identify this design. Show its data layout.
280 Factorial experiment—split-plot design
b. Write the mathematical model for this situation and determine the ANOVA
outline indicating tests to be made.
c. Recommend how many replications r you would run to make this an accept-
able experiment and explain why.
13.16 If, in Problem 13.15, only one oven was available but cakes of all three recipes
could be inserted in the oven when set at a given temperature, how would the
problem differ from the setup in Problem 13.15? Explain which of the two designs
you would prefer and why.
13.17 Running two random replications of a problem cited in Cochran and Cox
[Experimental Designs, 2nd edition, Wiley, 1957] on the effect of temperature and
mix on cake quality similar to Problem 13.15 resulted in the following data.
o Temperatures
OO
1 47 29 35 47 57 45
I 2 35 46 47 39 52 61
3 43 43 43 46 47 58
1 26 28 32 25 37 33
II 2 21 21 28 26 27 20
3 21 28 25 25 31 25
Outline the ANOVA table for this problem and show what tests can be run and
discuss these tests.
13.18 Analyze the data given in Problem 13.17 and state your conclusions.
13.19 What next steps would you recommend after your conclusions in Problem 13.18?
14
Factorial Experiment —
Confounding in Blocks
14.1 INTRODUCTION
As seen in Chapter 13, there are many situations in which randomization is
restricted. The split-plot design is an example of a restriction on randomiza-
tion where a main effect is confounded with blocks. Sometimes these restric-
tions are necessary because the complete factorial cannot be run in one day,
or in one environmental chamber. When such restrictions are imposed on the
experiment, a decision must be made as to what information may be sacrificed
and thus on what is to be confounded. A simple example may help to illustrate
this point.
A chemist was studying the disintegration of a chemical in solution. He
had a beaker of the mixture and wished to determine whether or not the depth
in the beaker would affect the amount of disintegration and whether material
from the center of the beaker would be different than material from near the
edge. He decided on a 2^ factorial experiment—one factor being depth, the
other radius. A cross section of the beaker (Figure 14.1) shows the four treat-
ment combinations to be considered.
In Figure 14.1 (1) represents radius zero, depth at lower level (near
bottom of beaker); a represents the radius at near maximum (near the edge of
the beaker) and a low depth level; b is radius zero, depth at a higher level;
and ab is both radius and depth at higher levels. In order to get a reading on
Figure 14.1
281
282 Factorial experiment—confounding in blocks
Plan
Dip I II III
If plan I of Table 14.1 is used, the blocks (dips) are confounded with the
radius effect, since all zero-radius readings are in dip 1 and all maximum
radius readings are in dip 2. In plan II the depth effect is confounded with
the dips, as low-level depth readings are both in dip 1 and high-level depth
readings are in dip 2. In plan III neither main effect is confounded, but the
interaction is confounded with the dips. To see that the interaction is con-
founded, recall the expression for the AB interaction in a 2^ factorial found
in Chapter 6.
AB = i[(I) — a — b nfi]
Note that the two treatment effects with the plus sign, (1) and ah, are both in
dip I, and the two with a minus sign, a and b, are in dip 2. Hence we cannot
distinguish between a block effect (dips) and the interaction effect (AB). In
most cases it is better to confound an interaction than to confound a main
effect. The hope is, of course, that there is no interaction, and that informa-
tion on the main effects can still be found from such an experiment.
as in the example above, he has only one interaction in the experiment and
decides that he can confound this interaction with blocks, the problem is
simply which treatment combinations to place in each block. One way to
accomplish this is to place in one block those treatment combinations with
a plus sign in the effect to be confounded, and those with a minus sign in the
other block. A more general method is necessary when the number of blocks
and the number of treatments increase.
First a defining contrast is set up. This is merely an expression stating
which effects are to be confounded with blocks. In the simple example
above, to confound AB, write AB as the defining contrast. Once the defining
contrast has been set up, several methods are available for determining which
treatment combinations will be placed in each block. One method was shown
above—place in one block the treatment combinations that have a plus
sign in AB and those with a minus sign in the other block. Another method
is to consider each treatment combination. Those that have an even number
of letters in common with the effect letters in the defining contrast go in one
block. Those that have an odd number of letters in common with the de-
fining contrast go into the other block. Here (1) has no letters in common
with AB or an even number. Type a has one letter in common with AB
(a and A) or an odd number, b also has one letter in common with AB. ab
has two letters in common with AB. Thus the block contents are those in
Figure 14.2.
AB
Block I Block II
Evens
Figure 14.2
One disadvantage of the two methods given above is that they are only
good on 2^ factorials. A more general method is attributed to Kempthorne
[13]. Consider the linear expression
A^X^ + A X 2 2 + ■ ■ ■ + A„X„ = L
L = 0 block 1 (1) ab
L = \ block 2 a
L = X, A X 2 A 2X3
and this can be used to decide which treatment combinations in a 3^ experi-
ment go into a common block.
The block containing the treatment combination (1) is called the princi-
pal block. The treatment combinations in this block are elements of a group
where the operation on the group is multiplication, modulus 2. The elements
of the other block or blocks may be generated by multiplying one element
in the new block by each element in the principal block. Multiplying elements
together within the principal block will generate more elements within the
principal block. This is best seen with a slightly more complex factorial, say
2^. If the highest order interaction is to be confounded and the eight treat-
ment combinations are to be placed into two blocks of four treatments each,
then confounding ABC gives
L, = X^ A X 2 + X3
Testing each of the eight treatment combinations gives
block 1 block 2
(1) a
ab b
ae c
be abc
L = 0 L = 1
Group theory can simplify this procedure. Determine two elements in the
principal block, for example, ab and be. Multiply these two to get ab^c = ac,
the fourth element. Generate block 2 by finding one element, say a. Multiply
this element by each element in the principal block, thus generating all the
elements of the second block as follows:
a ’ (1) = a
a • ab = a^b = b
a ■ ac = a^c = c
a ■ be = abc
The methods given earlier in this chapter can be used to determine the block
composition for a specific confounding scheme. If only one replication is
possible, one is run, and some of the higher order interaction terms must be
used as experimental error unless some independent measure of error is
available from previous data. This type of design is used mostly when there
are several factors involved (say four or more), some high-order interactions
may be confounded with blocks, and some others are yet available for an
error estimate. For example, consider a 2"^ factorial where only eight treat-
ment combinations can be run at one time. One possible confounding
286 Factorial experiment—confounding in blocks
scheme would be
and the analysis would be as shown in Table 14.2. The three-way interactions
may be pooled here to give 4 df for error, assuming that these interactions are
actually nonexistent. All main effects and first-order interactions could then
be tested with 1 and 4 df. After examining the results it might be possible to
pool some of the two-way interactions that have small mean squares with the
three-way interactions if more degrees of freedom are desired in the error
term.
If only four treatment combinations can be run in a block, one may
confound the 2^ factorial in four blocks of four treatment combinations each.
For four blocks 3 df must be confounded with blocks. If two interactions
are confounded, the product (modulus 2) of these two is also confounded,
since the product of the signs in two effects gives the signs for the product.
Thus, if AB and CD are confounded, so also will ABCD be confounded.
This scheme would confound two first-order interactions and one third order.
It might be better to confound two second-order interactions and one first
order:
ABC BCD AD
Table 14.2 2“^ Factorial in Two Blocks
Source df
A 1
B 1
C 1
D 1
AB 1
AC 1
AD 1
BC 1
BD 1
CD 1
ABC 1]
ABD 1
> 4 df for error
ACD 1
BCD ij
Blocks or ABCD 1
Total 15 df
14.3 Block confounding—no replication 287
+ X2 + X2
L2 = X2 + X2 + x^
Each treatment combination will give one of the following four pairs when
substituted into both and L = 00, 01, 10, 11. For example,
2
(l):Li =0 L2 = 0
Cl. L—— 1 B2 ~~ 0
h: Li = 1 L2 = 1
ab: Li = 2 = 0 L2 = 1
and so on. All treatment combinations with the same pair of L values will be
placed together in one block. Thus the 16 treatment combinations will be
assigned to four blocks as in Table 14.3.
Here the order of experimentation within each block is randomized and
the resulting responses are given for each treatment combination. The proper
treatment combinations for each block can be generated from the principal
block. If be and acd are in the principal block, then their product
the entries in the second block. In the same manner, the other two blocks may
be generated. This means that only a few readings in the principal block need
Table 14.3 Blocking a 2"^ Factorial in Four Blocks of Four
Confounding ABC, BCD, and AD
(1) = 82 a = 76 b = 19 ab = 85
be = 55 abc = 74 c = 71 ac = 84
acd = 81 cd = 72 abed = 89 bed = 84
abd — 88 bd = 73 ad = 79 d = m
L, = 0 Li = 1 Li = 1 Li = 0
L 2 = 0 Lj = 0 ~ T2 = 1 1^2=1
288 Factorial experiment—confounding in blocks
to be determined by the Lj, L2 values, and all the rest of the design can be
generated from these few treatment combinations. Tables are also available
for designs when certain effects are confounded [9].
This experiment is then a 2"^ factorial experiment. The design is a ran-
domized incomplete block design where blocks are confounded with the
interactions ABC, BCD, and AD. An example will illustrate the method.
Example 14.1 A systematic test was to be made on the effects of four
variables, each at two levels, on coil breakdown voltage. The variables were
A—firing furnace 1 or 3; B—firing temperature 1650°C or 1700°C; C—gas
humidification, no or yes; and D—coil outside diameter small (< 0.0300 inch)
or large (>0.0305 inch). Since only four coils could be tested in a given short
time interval, it was necessary to run four blocks of four coils for such testing.
Only one complete test of 2"*’ in four blocks of four was run and the results
(after dividing the voltages through by 10) are shown in Table 14.3, which
confounds ABC, BCD, and AD with the four blocks. For the analysis one
might use Yates’ method given in Chapter 6 to give Table 14.4.
In Table 14.4 the sums of squares in the last column correspond to the
effect for the treatment combination on the left. That is,
SS^ = 225.00 SSg = 0.25 and so on
Treatment
Combination Response (1) (2) (3) (4) SS
Total 1031.00
Note that the effects ABC, BCD, and AD are confounded with blocks. Their
total sum of squares is 42.25 + 156.25 + 1.00 = 199.50. If the above block
totals are computed, they yield
Source df SS MS
A 1 225.00 225.00
B 1 0.25 0.25
C 1 64.00 64.00
D 1 100.00 100.00
AB 1 56.25 56.25
AC 1 64.00 64.00
BC 1 12.25 12.25
BD 1 110.25 110.25
CD 1 121.00 121.00
Error or ABD, ACD, ABCD 3 78.50 26.17
Blocks or ABC, BCD, AD 3 199.50 66.50
Totals 15 1031.00
With only 1 and 3 df, none of the four main effects or the five two-way
interactions can be declared significant at the 5 percent significance level.
Since the B effect and the BC interaction are not significant even at the
25 percent level, we might wish to pool these with the error to increase the
error degrees of freedom. The resulting error sum of squares would then be
Complete Confounding
ABC
As seen in Section 14.2, the blocks would be
block 1 block 2
(1) a
ab b
ac c
be abc
If this whole experiment (2^ in two blocks of four each) can be replicated, say
three times, the layout might be
ac a c (1) ab c
(1) c abc ac (1) b
ab abc b be ac abc
be b a ab be a
14.4 Block confounding with replication 291
The confounding scheme in the above (ABC) is the same for all three replica-
tions, but the order of experimentation has been randomized within each
block. Also, the decision as to which block is to be run first in each replica-
tion is made at random. An analysis layout for this experiment appears in
Table 14.6.
Source df df
Replications 2
Blocks or ABC 1
Replications x block interactions 2 5 between plots
A 1
B 1
AB 1
C 1
AC 1
BC 1
Replications x all others 12 18 within plots
Totals 23 23
Here the replication interaction with all three main effects and their
interactions are usually taken as the error term for testing the important
effects. The replication effect and the block (or ABC) effect could be tested
against the replication x block interaction, but the degrees of freedom are
low and the power of such a test is poor. Such a design is quite powerful,
however, in testing the main effects A, B, and C and their first-order interac-
tions. However, it will give no clean information on the ABC interaction.
Partial Confounding
In the example just considered, we might be concerned with some test on the
ABC interaction. This could be found by confounding some interactions
other than ABC in some of the replications. We might use four replications
and confound ABC in the first one, AB in the second, AC in the third, and
BC in the fourth. Thus the four replications will yield full information on
A, B, and C but three-fourths information on AB, AC, BC, and ABC, since
we can compute an unconfounded interaction such as AB in three out of four
of the replications. The following layout shows the proper block entries for
each replication.
292 Factorial experiment—confounding in blocks
L = X, + X2 L = X^ A X2 L = Vi + V3 L = X2 A X2
A V3
The analysis layout would be that shown in Table 14.7.
The residual term with its 17 df can be explained as follows. In all four
replications the main effects A, B, and C can be isolated. This gives 3 x 1 df
for each replication by main-effect interaction, or a total of 9 df. The AB,
AC, BC, and ABC interactions can be isolated in only three out of the four
replications. Hence the replications by AC have 2 x 1 ^ 2 df, or a total of
8 df for all such interactions. This gives 9 + 8 = 17 df for effects by replica-
tions, which is usually taken as the error estimate. The blocks and replications
are separated out in the top of Table 14.7 only to reduce the experimental
error. The numerical analysis of problems such as this is carried out in the
same manner as given in earlier chapters.
Example 14.2 Coils were wound from each of two winding machines using
either of two wire stocks and their outside diameters were measured at two
Source df df
Replications 3
Blocks within replications 4 7 between plots
ABC (Replication I) 1
AB (Replication II) 1
AC (Replication III) 1
BC (Replication IV) 1
A 1
B 1
C 1
AB n only from
AC 1 replications
BC 1 where not
ABC ij confounded
Replications x all effects 17 24 within plots
Totals 31 31
14.4 Block confounding with replication 293
Table 14.8 Layout of Example 14.2 in Four Blocks of Four Confounding ABC
in All Replicates
replication I replication II
block 1 block 2 block 1 block 2
Source df SS MS F Prob.
Replications 3 409,719.59
Blocks or ABC 1 8,482.53
R X B interaction 3 515,407.85
7
Totals 31 10,064,614.22
294 Factorial experiment—confounding in blocks
Table 14.9 shows only positions on the coil as having a highly significant
effect on coil outside diameter.
Example 14.3 When coils from two other winding machines were to be
tested, it was decided to partially confound the interactions. Results gave
Table 14.10.
Table 14.10 Layout of Example 14.3 in Four Blocks of Four with Partial
Confounding
Analysis gives Table 14.11. In this table position on the coil is again
highly significant, but an AC interaction is also significant at the 5 percent
significance level.
In both problems the effect of A—machines—is approaching signifi-
cance. It may be that if more replications are run, this machine effect will
emerge as significant.
Confounding in 3^ Factorials
Source df SS MS F Probability
Replications 3 38,717.59
Blocks in
replications 4 401,060.13
7
Totals 31 9,002,296.97
00 10 20
11 21 01
22 02 12
Since there are three blocks, 2 df must be confounded. Here AB^ is con-
founded with blocks. If AB with its 2 df were confounded, L = X^ A X2,
296 Factorial experiment—confounding in blocks
00 10 20
12 22 02
21 01 11
If the data of Table 9.1 are used as an example, and AB^ is confounded, the
responses are
00 = 1 10 = -2 20-3
11 = 4 21 - 1 01-0
22-2 02 = 2 12 - -1
The block sums of squares are determined from the three block totals of
7, 1, and 2:
7^ + + 2^ (10)^
SS block 6.89
3
Source df SS MS
A 2 4.22 2.11
B 2 1.56 0.78
AB 2 16.22 8.11
Blocks or AB^ 2 6.89 3.44
Totals 8 28.89
14.4 Block confounding with replication 297
The analysis of data for this design would yield Table 14.13.
This is a useful design, since we can retrieve all main effects {A, B, C)
and all two-way interactions, if we are willing to pool the 6 df in ABC as
error. The determination of sums of squares, and so forth, is the same as for
a complete 3^ given in Example 9.1. Confounding other parts of the ABC
interaction would, of course, yield different blocking arrangements, although
the outline of the analysis would be the same. In practice, different blocking
arrangements might yield different responses.
More complex confounding schemes can be found in tables such as in
Chapter 9 of Davies [9].
For practice, consider confounding a 3^ in nine blocks of three treatment
combinations each. Here 8 df must be confounded with blocks. One con-
founding scheme might be
AB^C^ AB BC^ AC
Source df
A 2
B 2
AB 4
C 2
AC 4
BC 4
Error or ABC, ABC^, AB^C^ 6
Blocks or AB^C 2
Total 26
298 Factorial experiment—confounding in blocks
Note that
{AB^C^){AB) = A^C^ = AC
and
iAB){BC^) = AB^C^
These are not all independent. In fact, only two of the expressions are. For
this reason we need only two expressions for the L values;
L, = A 2X2 + 2X3
L2 = X,A X2
as these two will yield nine pairs of numbers—one pair for each block. First
determine the principal block, where both and L2 are zero. One treatment
combination is obviously 000. Another is 211, as
Li = 1(2) + 2(1) + 2(1) = 6 = 0
L2 = 1(2) + 1(1) = 3 = 0
A third is 122, as
Li = 1(1) + 2(2) + 2(2) = 9 = 0
L2 = 1(1) + 1(2) = 3 = 0
The other eight blocks can now be generated from this principal block,
giving
block 1 2 3 4 5 6 7 8 9
Source df
A 2
B 2
C 2
AB^ 2
AC^ 2
BC 2
ABC 2]
AB^C 2 > 6 df for error
ABC^ 2)
Blocks or AB, BC\ AC, AB^C^ 8
Total 26
14.5 Summary 299
14.5 SUMMARY
a. Complete a. Factorial
^ijk = /^ + + d,- ANOVA
+ Bj + AB^j +
b. Incomplete, confounding: b.
i. Main effect—split-plot i. Split-plot
Tjk = + Bi + Aj + RAij ANOVA
'' V ^
whole plot
+ + ^^ik + d-BjTc + V
RAB^j^
^
split-plot
ii. Interactions in 2^ and 3^ ii.
(1) Several replications (1) Factorial
Tjkq ^ y. + Ri + ANOVA
+ Rf^ij + A,^ + Bq with
replications
Rj and
blocks l^j or
confounded
interaction
(2) One replication only (2) Factorial
Tjk = B + Pi + ANOVA
+ ^k + .. . with blocks
PjOV
confounded
interaction
300 Factorial experiment—confounding in blocks
PROBLEMS
14.1 Assuming that a 2^ factorial such as the problem of Chapters 1, 5, and 6 involving
two tool types, two bevel angles, and two types of cut cannot all be run on one
lathe in one time period but must be divided into four treatment combinations
per lathe, set up a confounding scheme for confounding the AC interaction in
these two blocks of four. Assuming the results of one replication give (1) = 5,
u = 0, h = 4, uh = 2, c = —3,ac = 0,bc= —1, abc = — 2, show that the
scheme you have set up does indeed confound AC with blocks.
14.2 Assuming three replications of the 2^ experiment in Problem 14.1 can be run but
each one must be run in two blocks of four, set up a scheme for the complete
confounding of ABC in all three replicates and show its ANOVA layout.
14.3 Repeat Problem 14.2 using partial confounding of AB, AC, and BC.
14.4 Data involving four replications of the 2^ factorial described in the foregoing
problems and in Section 14.3 gave the following numerical results.
replication I replication II
confound ABC AB
(1) = 5 0 = 0 (1) = -2 h = 9
ac = 0 c = -3 he = —1 e = -3
o
abc = —2 he = —1 0 = —6
II
b = 4 ab = 2 abc = —4 ac — —4
(1) = 2, a = 1, be = 3, abc = 4
c = 1, ac = 4, b = 6, ab = 0
By examining the variation between the two days and your knowledge of a 2-^
experiment, hnd out which term in the model (main effect or interaction) is
confounded with days.
Problems 301
14.6 For the data of Problem 6.6 consider running the 16 treatment combinations in
four blocks of four (two observations per treatment). Confound ACD, BCD, and
AB and then determine the block compositions.
14.7 From the results of Problem 6.6 show that the blocks are confounded with the
interactions in Problem 14.6.
14.8 For Problem 9.11 consider this experiment as run in three blocks of nine with
two replications for each treatment combination. Confound ABC (2 df) where
A is surface thickness, B is base thickness, and C is subbase thickness, and deter-
mine the design. Also use the numerical results of Problem 9.13 to show that
ABC is indeed confounded in your design.
14.9 Set up a scheme for confounding a 2"^ factorial in two blocks of eight each,
confounding ABD with blocks.
14.10 Repeat Problem 14.9 and confound BCD.
14.11 Work out a confounding scheme for a 2^ factorial, confounding in four blocks
of eight each. Show the outline of an ANOVA table.
14.12 Repeat Problem 14.11 in four replications confounding different interactions in
each replication. Show the ANOVA table outline.
14.13 Work out the confounding of a 3‘^ factorial in three blocks of 27 each.
14.14 Repeat Problem 14.13 in nine blocks of nine each.
14.15 In Example 14.2, since ABC is completely confounded, the analysis could be
easily run on a computer or by using Yates’ method on the A, B, C sections and
general methods for replications and blocks. Verify the results given in Table 14.9.
14.16 Since Example 14.3 has partial confounding, it may be difficult for the computer.
However, a Yates method can be used if care is taken as to how it is applied.
Try to verify Table 14.11 results using a modified Yates procedure.
14.17 A systematic test was made to determine the effects on coil breakdown voltage
of the following six factors, each at two levels as indicated:
1. Firing furnace Number 1 or 3
2. Firing temperature 1650°C or 1700°C
3. Gas humidification No or yes
4. Coil outer diameter Below 0.0300 inch or above 0.0305 inch
5. Artificial chipping No or yes
6. Sleeve Number 1 or 2
All 64 experiments could not be performed under the same conditions so the
whole experiment was run in eight subgroups of eight experiments. Set up a
reasonable confounding scheme for this problem and outline its ANOVA.
14.18 An experiment was run on the effects of several annealing variables on the
magnetic characteristic of a metal. The following factors were to be considered.
1. Temperature 1375°, 1450°, 1525°, 1600°
2. Time in hours 2, 4
302 Factorial experiment—confounding in blocks
Position
1 2 3
1 1 0 0 0 0 2
2 -3 2 -1 -2 -2 -2
3 1 2 0 3 0 4
4 0 0 -2 1 0 1
5 -3 1 0 2 -2 0
6 -2 1 -3 2 0 0
7 3 3 3 3 1 -1
8 1 1 0 5 1 2
14.23 Consider a 2^ factorial experiment with all fixed levels to be run in four blocks
of eight treatment combinations each. The whole experiment can be replicated
in a total of three replications. The following interactions are to be confounded
in each replication.
15.1 INTRODUCTION
As the number of factors to be considered in a factorial experiment increases,
the number of treatment combinations increases very rapidly. This can be
seen with a 2^ factorial where / = 5 requires 32 experiments for one replica-
tion, / = 6 requires 64, / = 7 requires 128, and so on. Along with this
increase in the amount of experimentation comes an increase in the number
of high-order interactions. Some of these high-order interactions may be
used as error, as those above second order (three way) would be difficult to
explain if found significant. Table 15.1 gives some idea of the number of
main effects, first-order, second-order,. . ., interactions that can be recovered
if a complete 2^ factorial can be run.
Considering / = 7, there will be 7 df for the seven main effects, 21 df for
the 21 first-order interactions, 35 df for the 35 second-order interactions,
leaving
35 21 + 7 -f 1 = 64 df
for an error estimate, assuming no blocking. Even if this experiment were
confounded in blocks, there is still a large number of degrees of freedom for
the error estimate. In such cases, it may not be economical to run a whole
replicate of 128 observations. Nearly as much information can be gleaned
Order of Interaction
Main
/ 2^ Effect 1st 2nd 3rd 4th 5th 6th 7th
5 32 5 10 10 5 1
6 64 6 15 20 15 6 1
7 128 7 21 35 35 21 7 1
8 256 8 28 56 70 56 28 8 1
303
304 Fractional replication
15.2 ALIASES
To see how a fractional replication is run, consider a simple case. Three
factors are of interest, each at two levels. However, the experimenter cannot
afford 2^ = 8 experiments, but will, however, settle for four. This suggests a
one-half replicate of a 2^ factorial. Suppose ABC is confounded with blocks.
The two blocks are then
/ = ABC
block 1 (1) ab be ac
block 2 a b c abc
A coin is flipped and the decision is made to run only block 2. What informa-
tion can be gleaned from block 2 and what information is lost when only
half of the experiment is run?
Referring to Table 6.7, which shows the proper coefficients (+1 or — 1)
for the treatment combinations in a 2^ factorial to give the effects desired,
box only those treatment combinations in block 2 that are to be run (Table
15.2).
Note, in the boxed area, that ABC has all plus signs in block 2, which is
a result of confounding blocks with ABC. Note also that the effect of A is
/l = +a — h — c A- ahc
and
BC = Aa — h — c A abc
so that we cannot distinguish between A and BC in block 2. Two or more
effects that have the same numerical value are called aliases. We cannot tell
them apart. B and AC likewise are aliases, as are C and AB. Because of this
confounding when only a fraction of the experiment is run, we must check
the aliases and be reasonably sure they are not both present if such a design
is to be of value. An analysis of this one-half replication of a 2^ would be as
shown in Table 15.3.
15.2 Aliases 305
Effect
Treatment
Combination A B AB C AC BC
(1) —
+ —
+ -f-
a + — — — — + -f
b —
+ — — + — +
ab -f -f + _ _
c — — + + — — A
ac -f - — -I- -1- -
be - + - + - +
abc + + + -f + + -f
Source df
A (or BC) 1
B (or AC) 1
C (or AB) 1
Total 3
The dehnition given above still holds where one effect is the alias of another
if they have the same numerical value, or value regardless of sign.
A quick way to find the aliases of an effect in a fractional replication of a
2^ factorial experiment is to multiply the effect by the terms in the defining
306 Fractional replication
contrast, modulus 2. The results will be aliases of the original effect. In the
example above,
/ = ABC
The alias of A is
A{ABC) = A^BC = BC
The alias of B is
BiABC) = AB^C = AC
The alias of C is
C{ABC) = ABC^ = AB
This simple rule works for any fractional replication for a 2^ factorial. It
works also with a slight modification for a 3^ factorial run as a fractional
replication.
Source df
which confounds 3 df with the four blocks. In this design, if only one of the
four blocks of 32 observations is run, each effect has three aliases. These are
Source df
Total 31
This is quite a formidable list of aliases; but when only one block of 32 is run,
there are 31 df within the block. The above list accounts for these 31 df. If,
in this design, all second-order (three-way) and higher interactions can be
considered negligible, the main effects are all clear of first-order interactions.
Three first-order interactions {AB, AF, and AG) are each aliased with another
first-order interaction (FG, FG, and BF). If the choice of factors A, B, F, and
G can be made with the assurance that the above interactions are either
negligible or not of sufficient interest to be tested, these 3 df can be either
pooled with error or left out of the analysis. The remaining 15 first-order
interactions are all clear of other interactions except second order or higher.
There are also 6 df left over for error involving only second-order interactions
or higher. An analysis might be that shown in Table 15.5.
Here tests can be made with 1 and 6 df or 1 and 9 df if the last two lines
can be pooled for error. This is a fairly good design when it is necessary to
run only a one-fourth replication of a 2^ factorial.
Example 15.1 In an experiment on slicing yield from pork bellies, four
factors were considered as possible sources of significant variation. These
were A—press effect; B—molding temperature; C—slicing temperature;
D—side of the hog. If each of these four factors were set at two levels, it
would require 2"^ = 16 experiments to run one complete replication. If only
eight bellies could be handled at one time, it was decided to run a one-half
replicate of this 2"’’ factorial and confound ABCD with blocks. If / = ABCD,
the blocks are
block I (1) ah cd abed ae be ad bd
Now if only block I is run. Table 15.6 shows the data where entries are for
the treatment combinations from block I only.
The analysis could be handled by a Yates’ method using only the one-half
replication data available as shown in Table 15.7. Note that the numerical
alias of A is A(ABCD) = BCD, of F = ACD, etc. (shown by the arrows).
The analysis follows in Table 15.8.
Table 15.6 Slicing Yields in Percent from Pork Belly Experiment (Example 15.1)
Press {A)
Normal High
Molding Temperature (B) Molding Temperature
Slicing
Temperature Side 23° 40° 23° 40°
Treatment
Combination Y (1) (2) (3) (4) ss
(1) 95.29 95.29 181.87 357.36 723.IN
a 86.58 175.49 365.75 -0.69K\ 0.06
b 88.70 185.83 -6.80 -18.47,^\\ 42.64
ab 86.58 86.79 179.92 6.29 0.47,<X^ 0.03
e 96.45 -8.71 -10.62 -12.29 \ 18.88
ae 88.70 89.38 1.91 -7.85 2.77r\\\ \\ 0.96
be 86.79 90.35 7.07 6.38 13.09A\\\ \\ 21.42
abe 89.57 -0.78 -5.91 8.390 \ ' 8.78
d -95.29 -8.71 -6.38 8.392J j
ad 96.45 86.58 -1.91 -5.91 13.09^ J
bd 89.38 88.70 -7.07 10.62 2.71'^///1
abd -86.79 -0.78 -7.85 -12.29'V^
ed 90.35 96.45 181.87 6.80
aed -89.38 -175.49 6.29 -\ZAlW /
bed -90.35 -185.83 -357.36 -0.69'^
abed 89.57 89.57 179.92 365.75 723.1 r
Source df SS MS F Prob.
309
310 Fractional replication
This experiment is not too good because the two way interactions are
hopelessly confounded and the degrees of freedom on error are so small.
Nothing shows significance but the analysis concept is seen by this example.
3^ Factorials
The simplest 3^ factorial is a 3^ requiring nine experiments for a complete
factorial. If only a fraction of these can be run, we might consider a one-third
replication of a 3^ factorial. First we confound either AB or AB^ with the
three blocks. Confounding AB^ gives
/ = AB^
L = + 23^2
and the three blocks are
L = 0 L=1 L = 2
00 10 20
11 21 01
22 02 12
Source df
Total 8
Factor A
Factor B Factor C 0 1 2
2 0 mo. m 220
1 021 m 221
2 122 222
Factor A
Factor B 0 1 2
0 2 0 1
1 0 1 2
2 1 2 0
15.3 Fractional replications 313
where A, B, and C are taken from the treatment effect. Both models look
alike but derive from different types of experiments. In both cases the
assumption is made that there is no interaction among the factors. This is
more reasonable when the factors represent only randomization restrictions
such as blocks and positions, rather than when each of the three factors is
of vital concern to the experimenter.
By choosing each of the other two blocks in the confounding of ABC^
(L = 0 or L = 2), two other Latin squares are obtained. If another con-
founding scheme is chosen, such as AB^C, ABC, or AB^C^, more and
different Latin squares can be generated. In fact, with four parts of the
three-way interaction and three blocks per confounding scheme, 12 distinct
Latin squares may be generated. This shows how it is possible to select a
Latin square at random for a particular design.
It cannot be overstressed that we must be quite sure that no interaction
exists before using these designs on a 3^ factorial. It is not enough to say
that there is no interest in the interactions, because, unfortunately, the
interactions are badly confounded with main effects as aliases.
Example 15.2 Torque after preheat in pounds was to be measured on some
rubber material with factors A—temperature at 145°C, 155°C, and 165°C;
B—mix I, II, and III; and C—laboratory 1, 2, and 3. The 3^ = 27 experi-
ments seemed too costly, so a one-third replicate was run. Confounding
AB^C gives L = X^ 5- 2X2 + X^ and block composition as follows:
If, now, block L = 1 only is run, the numerical results are as given in
Table 15.12 and the analysis as in Table 15.13.
Temperature [A)
Source df SS MS
Totals 8 85.72
Because of the tiny error term all main effects are highly significant
even with only 2 and 2 degrees of freedom.
15.4 SUMMARY
Continuing the summary at the end of Chapter 14 under incomplete ran-
domized blocks gives
PROBLEMS
15.1 Assuming that not all of Problem 14.1 can be run, determine the aliases if a one-
half replication is run. Analyze the data for the principal block only.
15.2 Assuming only a one-fourth replication can be run in Problem 14.6, determine
the aliases.
15.3 For Problem 14.8 use a one-third replication and determine the aliases.
15.4 Run an analysis on the block in Problem 15.3 that contains treatment combina-
tion 022.
15.5 What would be the aliases in Problem 14.9 if a one-half replication were run?
15.6 Determine the aliases in a one-fourth replication of Problem 14.11.
15.7 Determine the aliases in Problem 14.13 where a one-third replication is run.
15.8 If a one-eighth replication of Problem 14.17 is run, what are the aliases? Comment
on its ANOVA table.
15.9 What are the aliases and the ANOVA outline if only one block of Problem 14.18
is run?
15.10 If data are taken from one furnace only in Problem 14.21, what are the aliases
and the ANOVA outline?
15.11 What are the numerical results and conclusions from just one furnace of Prob-
lem 14.22?
15.12 The experiment of Example 15.1 was later continued and the second block was
now run with results as follows:
Press {A)
Normal High
L b = 88.08 a = 89.26
23°
R d = 90.87 abd = 93.16
L c = 85.95 abc = 85.68
30°
R bed — 83.65 acd = 95.25
b. Outline the ANOVA for just one run of this experiment. Consider all factors
hxed and note what you might use as an error term.
c. If only one block of this experiment can be run, explain what complications
this makes in the experiment and how the ANOVA table must be altered.
15.15 Consider a 2^ factorial experiment with all hxed levels to be run in four blocks
of eight treatment combinations each. This whole experiment can be replicated
in a total of three replications. The following interactions are to be confounded
in each replication:
Replication I — ABC, CDE, ABDE
Replication II — BCDE, ACD, ABE
Replication III — ABCD, BCE, ADE
a. Determine the block composition for the principal block in replication I.
b. If only a one-fourth replication were run of replication II, what would be the
aliases of BC in this block?
c. Determine the number of degrees of freedom for the replication by three-way
interactions in this problem.
15.16 Consider an experiment designed to study the effect of six factors—A, B, C, D,
E, and E—each at two levels.
a. Determine at least six entries in the principal block if this experiment is to be
run in two blocks of 32 and the highest order interaction is to be confounded.
b. If only one of the two blocks can be run, what are the aliases of the effects
A, BC, and /45C?
c. Outline the ANOVA table for part (b) and indicate what effects might be
tested and how they would be tested.
15.17 In determining which of four factors A, B, C, or D might affect the underpuffing
of cereal in a plant, the following one-fourth replicate was run.
Bi B, B2
96.6 125.7
c,1
D, 43.5 22.4
D, 14.1 9.5
c ■)
Di 28.8 52.5
<60 >62
C Sulfur, percent
2 3
>59
oo
1 0 0
>2845
>10
Tundish temperature, °F
5 0
Casting speed, 1 pm
oo
1 6 0
>10
6 3
>59
F
E
oo
1 2 0
D
ly-)
<2840
>10
7 0
<54
oo
1 1 0
16.1 INTRODUCTION
Several techniques have been developed, based on the general design princi-
ples presented in the first 15 chapters of this book and on other well-known
statistical methods. No attempt will be made to discuss these in detail, as
they are presented very well in the references. It is hoped that a background
in experimental design as given in the first 15 chapters will make it possible
for the experimenter to read and understand the references.
The methods to be discussed are covariance analysis, response surface
experimentation, evolutionary operation, analysis of attribute data, incom-
plete block design, and Youden squares.
Occasionally, when a study is being made of the effect of one or more factors
on some response variable, say T, there is another variable, or variables,
that vary along with Y. It is often not possible to control this other variable
(or variables) at some constant level throughout the experiment, but the
variable can be measured along with the response variable. This variable is
referred to as a concomitant variable X as it “runs along with” the response
variable Y. In order, then, to assess the effect of the treatments on Y, one
should first attempt to remove the effects of this concomitant variable X.
This technique of removing the effect of X (or several X’s) on Y and then
analyzing the residuals for the effect of the treatments on Y is called co-
variance analysis.
Several examples of the use of this technique might be cited. If one
wishes to study the effect on student achievement of several teaching methods,
it is customary to use the difference between a pretest and a posttest score
as the response variable. It may be, however, that the gain Y is also affected
318
16.2 Covariance analysis 319
by the student’s pretest score X as gain, and pretest scores may be correlated.
Covariance analysis will provide for an adjustment in the gains due to
differing pretest scores assuming some type of regression of gain on pretest
scores. The advantage of using this technique is that one is not limited to
running an experiment on only those pupils who have approximately the
same pretest scores or matching pupils with the same pretest scores and
randomly assigning them to control and experimental groups. Covariance
analysis has the effect of providing “handicaps” as if each student had the
same pretest scores while actually letting the pretest scores vary along with
the gains.
This technique has often been used to adjust weight gains in animals by
their original weights in order to assess the effect of certain feed treatments
on these gains. Many industrial examples might be cited such as the one
given below where original weight of a bracket may affect the weight of
plating applied to the bracket by different methods.
Snedecor [22] gives several examples of the use of covariance analysis.
Ostle [19] is an excellent reference on the extension of covariance analysis
to many different experimental designs. The Biometrics article of 1957 [5]
devotes most of its pages to a discussion of many of the fine points of covari-
ance analysis. The problem discussed below will illustrate the technique for
the case of a single-factor experiment run in a completely randomized
design. Since the covariance technique represents a marriage between re-
gression and the analysis of variance, the methods of Chapter 3 and Chapter 7
will be used.
Example 16.1 Several steel brackets were sent to three different vendors to
be zinc plated. The chief concern in this process is the thickness of the zinc
plating and whether or not there was any difference in this thickness between
the three vendors. Data on this thickness in hundred thousandths of an inch
(10“^) for four brackets plated by the three vendors are given in Table 16.1.
Assuming a single-factor experiment and a completely randomized
design, an analysis of variance model for this experiment would be
+ C + (16.1)
Table 16.1 Plating Thickness in 10”^ Inches
from Three Vendors
Vendor
A B C
40 25 27
38 32 24
30 13 20
47 35 13
320 Miscellaneous topics
Source df SS MS
Totals 11 1208.7
where Y^j is the observed thickness of the ith bracket from the jth vendor,
ji is 3. common effect, TJ represents the vendor effect, and Sij represents the
random error. Considering vendors as a fixed factor and random errors as
normally and independently distributed with mean zero and common vari-
ance (jg, an ANOVA table was compiled (Table 16.2).
The F statistic equals 5.51 and is significant at the 5 percent significance
level with 2 and 9 df One might conclude, therefore, that there is a real
difference in average plating thickness among these three vendors, and
steps should be taken to select the most desirable vendor.
During a discussion of these results, it was pointed out that some of
this difference among vendors might be due to unequal thickness in the
brackets before plating. In fact, there might be a correlation between the
thickness of the brackets before plating and the thickness of the plating. To
see whether or not such an idea is at all reasonable, a scatterplot was made
on the thickness of the bracket before plating X and the thickness of the
zinc plating Y. Results are shown in Figure 16.1.
A glance at this scattergram would lead one to suspect that there is a
positive correlation between the thickness of the bracket and the plating
50
40
30
Y
20
10
0 30 60 90 120 150
V
Covariance
where P is the true linear regression coefficient (or slope) between Y and X
over all the data; X is the mean of the X values. In such a model it is assumed
that a measure of variable X can be made on each unit along with a corre-
sponding measure of Y. For a covariance analysis models (16.1) and (16.2)
can be combined to give the covariance model
In this model the error term eij should be smaller than in model (16.1)
because of the removal of the effect of the covariate X^j.
To determine how to adjust the analysis of variance to provide for the
removal of this covariate, consider the regression model (16.2) in deviation
form
y = bx + e (16.4)
where x and y are now deviations from their respective means (x = X^j — X,
y = Yij — Y) and b is the sample slope. By the method of least squares it
will be recalled that
where summations are over all points. The sum of the squares of the errors
of estimate, which have been minimized, is now, from Equation (16.4),
e = y — bx
y = y (y - hx}^
= - Ib^xy +
Substituting for b
Z xy
322 Miscellaneous topics
and
Z xy
Z xy + 'Z z
z^^
= z/- (Zz xy)^ (16.5)
z x^
(II^
N
iZll
N
(Z
N
Data on both variables appear with the appropriate totals in Table 16.3.
From Table 16.3 the sums of squares and cross products for the total
data are^
2
r,, = (110)^ + (75)^ + ••• + (59)^ -
(944)2 = 9240.7
12
(944)(344)
T,, = (110)(40) + (75)(38) + • • • + (59)(13) - = 2332.7
12
The between-vendors sums of squares and cross products are computed by
' Here T, V, and E are used to denote sum of squares and cross products for totals, vendors,
and error respectively.
16.2 Covariance analysis 323
Vendor
A B C Total
X y X y X y X y
no 40 60 25 62 27
75 38 75 32 90 24
93 30 38 13 45 20
97 47 140 35 59 13
(2332.7)2
= 1208.7 619.8
9240.7
adjusted = Eyy —
E XX
(1270.5)^
= 543.5 - = 327.4
7469.5
324 Miscellaneous topics
SS and Products
Adjusted
Source df IV I^^ IV df MS
146.2
3.57
40.9
with 2 and 8 df. This is now not significant at the 5 percent significance level.
This means that after removing the effect of bracket thickness, the vendors
no longer differ in the average plating thickness on the brackets.
Table 16.4 should have a word or two of explanation. The degrees of
freedom on the adjusted sums of squares are reduced by 2 instead of 1 as
estimates of both the mean and the slope are necessary in their computation.
Adjustments are made here on the totals and the within sums of squares
rather than on the between sum of squares as we are interested in making
the adjustment based on an overall slope of 7 on AT and a within-group
average slope of T on 2f as explained in more detail later.
In this analysis three different types of regression can be identified: the
“overah” regression of all the T’s on all the X’s, the within-vendors regres-
sion, and the regression of the three Y means on the three X means. This last
regression could be quite different from the other two and is of little interest
since we are attempting to adjust the T’s and the X’s within the vendors. An
estimate of this “average within vendor” slope is
1270.5
0.17
“ 7469.5
If this slope or regression coefficient is used to adjust the observed Y means
for the effect of X on T, one finds
adjusted Y j = Y j — b{X j — )
16.2 Covariance analysis 325
For vendor A,
a = vendor A
A - mean for vendor A
b = vendor B
B = mean for vendor B
c = vendor C
C = mean for vendor C
much steeper. Hence adjustments are made on the basis of the overall slope
and the “average” or “pooled” within-groups slope.
One of the problems in covariance analysis is that there are many
assumptions made in its application. These should be examined in some
detail.
First, there are the usual assumptions of analysis of variance on the
dependent variable Y: an additive model, normally and independently dis-
tributed error, and homogeneity of variance within the groups. This last
assumption is often tested with Barlett’s or a similar test. In this problem,
since the ranges of the three vendor readings are 17, 22, and 14, the assump-
tion seems tenable.
In covariance analysis it is further assumed that the regression is linear,
that the slope is not zero (covariance was necessary), that the regression co-
efficients within each group are homogeneous so that the “average” or
“pooled” within-groups regression can be used on all groups, and finally
that the independent variable X is not affected by the treatments given to the
groups.
The linearity assumption may be tested if the experiment is planned in
such a way that there is more than one Y observation for each X. Since this
was not done here, a look at the scattergram will have to suffice for this
linearity assumption.
To test the hypothesis that the true slope p in Equation (16.3) is zero,
consider whether or not the reduction in the error sum of squares is significant
when compared with the error sum of squares. By an F test the ratio would be
adjusted Eyy/{N — k — 1)
^ _ 216.1 _ 216.1 _
“ 327.4/8 ~ ~W.9 ~ ■
As the 5 percent F is 5.32, this result is nearly significant and certainly casts
considerable doubt on the hypothesis that p = 0. Since the results after
covariance adjustment were not significant as compared with the results
before adjustment, it seems reasonable to conclude that covariance was
necessary.
To test the hypothesis that all three regression coefficients are equal, let
us compute each sample coefficient and adjust the sum of squares within each
group by its own regression coefficient.
For vendor A,
Within group A,
14,599 - 58,125/4 ^
35,783 - 140,625/4
For vendor 5,
, 9294 - 32,865/4
® ~ 30,269 - 24,492.25/4 “ '
Within Adjusted
Vendor df b df
l, N-2k ~
and
(321A - 319.8)/2 3.8
^ 2,6 ~~
= <1
319.8/6 53.3
_ 1771.2/2 _ 885.6
1.07
“ 7469.5/9 “ 829.9
PhUosophy
(a)
(d)
Figure 16.3 Some typical response surfaces in two dimensions: (a) mound;
(b) depression; (c) rising ridge; (d) saddle.
For the case of two independent variables such as temperature and time
^2, the yield 7 of a chemical process can be expressed as
Y = fix
This surface can be plotted in three dimensions, with X^ on the abscissa,
X on the ordinate, and Y perpendicular to the X^X plane. If the values
2 2
of X^ and X that yield the same Y are connected, we can picture the surface
2
with a series of equal-yield lines, or contours. These are similar to the contours
of equal height on topographic maps and the isobars on weather maps.
Some response surfaces might be of the types in Figure 16.3. Two excellent
references for understanding response-surface experimentation are [9] and
[12].
The Twofold Problem
next experiment toward the optimal point on the underlying response surface;
(2) having located the optimum or near optimum of the surface, to determine
the equation of the response surface in an area near this optimum point.
One method of experimentation that seeks the optimal point of the
response surface might be the traditional one-factor-at-a-time method. As
shown in Figure 16.3(a), if X is fixed and
2 is varied, we find the X^ optimal
(or near optimal) value of response Y at the fixed value of X2. Having found
this X^ value, experiments can now be run at this fixed X^, and the X for 2
optimal response can be found. In the case of the mound in Figure 16.3(a),
this method would lead eventually to the peak of the mound or near it.
However, this same method, when applied to a surface such as the rising
ridge in Figure 16.3(c), fails to lead to the maximum point on the response
surface. In experimental work the type of response surface is usually un-
known, so that a better method is necessary if the optimum set of conditions
is to be found for any surface.
The method developed by those who have worked in this area is called
the path of steepest ascent method. The idea here is to run a simple experiment
over a small area of the response surface where, for all practical purposes,
the surface may be regarded as a plane. We then determine the equation of
this plane and from it the direction we should take from this experiment in
order to move toward the optimum of the surface. Since the next experiment
should be in a direction in which we hope to scale the height the fastest, this
is referred to as the path of steepest ascent. This technique does not determine
how far away from the original experiment succeeding sequential experiments
should be run, but it does indicate to the experimenter the direction along
which the next experiment should be performed. A simple example will
illustrate this method.
To determine the equation of the response surface, several special
experimental designs have been developed that attempt to approximate this
equation using the smallest number of experiments possible. In two dimen-
sions the simplest surface is a plane given by
Y = BQXQ + B^X^ + B X
2 2 + 8 (16.6)
where Y is the observed response, XQ is taken as unity, and estimates of the
JB’S are to be determined by the method of least squares, which minimizes the
sum of the squares of the errors e. Such an equation is referred to as a first-
order equation, since the power on each independent variable is unity.
If there is some evidence that the surface is not planar, a second-order
equation in two dimensions may be a more suitable model
Y = BQXQ + B^X^ + B X 2 2 B^^Xf + B^2^i^2 T -^22-^2 T ^ (16.7)
Here the X^X term represents an interaction between the two variables X^
2
and X -
2
16.3 Response-surface experimentation 331
As there are three parameters to be estimated, BQ, B^, and B , at least three 2
concentrations were chosen, 1 and 1:1 (ratio of filler to resin), and two
positions, 1 inch and 2 inches from a reference point, with responses Y, the
thickness of the material in lO""'' inches, as shown in Figure 16.4.
To determine the equation of the best fitting plane for these four points,
consider the error in the prediction equation
5.4 6.9
+1 2 in.
Position X2
-1 1 in.
1;1
Concentration
-1 +1
where the summation is over all points given in the design. To find the 5’s,
we differentiate this expression with respect to each parameter and set these
three expressions equal to zero. This provides three least squares normal
equations, which can be solved for the best estimates of the J5’s. These esti-
mates are designated as fi’s.
Differentiating the expression above gives
= -2 X (r - = 0
= -2 ^ (y - BoX„ - BiXi - B X )X 2 2 2 = 0
Y Xo Xi X2
7.0 1 -1 -1
5.4 1 -1 1
7.1 1 1 -1
6.9 1 1 1
XQ is always taken as unity. For the experimental variables X-^ and X2, the
responses can be recorded as in Table 16.6. These experimental variables are
also indicated on Figure 16.4.
An examination of the data as presented in Table 16.6 shows that XQ,
X^, and X2 are all orthogonal to each other as
= X X2 = 0 and X ^1^2 = 0
X XQY = bQU + 61 • 0 + ^2 • 0
Y,X,Y = ho-o + h.-Y^xi + b^-o (16.11)
X ^2^ = 60 • 0 4- • 0 + 62 X ^2
Solving gives
^0 = i:XoY/n
(16.12)
b2 = X^2 5"/X ^2
For this problem
1.6
61 = = 0.40
4
bi = = -0.45
4
To determine the sum of squares due to each of these terms in the model,
we can use the sum of squares due to h,’s
334 Miscellaneous topics
Here
= (6.60)(26.4) = 174.24
= (0.40)(1.6) = 0.64
SS,^ = (-0.45)(-1.8) = 0.81
each carrying 1 df. The ANOVA table is shown as Table 16.7.
Source df SS
bo 1 174.24
hi 1 0.64
bi 1 0.81
Residual 1 0.49
Totals 4 176.18
Here, all 4 df are shown, since the ho term represents the degree of
freedom usually associated with the mean, that is, the correction term in
many examples. The total sum of squares is ^ of the responses and the
residual is what is left over. With this 1-df residual, no good test is available
on the significance of each term in the model, nor is there any way to assess
the adequacy of the planar model to describe the surface.
To decide on the direction for the next experiment, plot contours of
equal response using the equation of the plane determined above
6.60 - y + 0.40X1
045
If y =
1.10 + 0.40X1
045
when
Xi = -1 X 2 = 1.56
Xi = 1 X2 = 3.33
0.60 + 0.40X1
0.45
16.3 Response-surface experimentation 335
when
X, = -1 X2 = 0.44
X^ = 1 ^2 = 2.22
If y -
0.10 + 0.40Xi
~ 045
when
X^ = -1 X2= -0.67
= 1 X2 = 1.11
-0.40 + 0.40X1
045
when
Xi = -1 X2 = -1.67
Xi = 1 X2 = 0
If y =
_ -0.90 + 0.40X1
^ 0.45
when
Xi = -1 X2 = -2.89
Xi = 1 X2 = -1.11
Plotting these five contours on the original diagram gives the pattern shown
in Figure 16.5.
By moving in a direction normal to these contours and “up” the surface,
we can anticipate larger values of response until the peak is reached. To
^2
decide on a possible set of conditions for the next experiment, consider the
equation of a normal to these contours through the point (0, 0). The con-
tours are
6.60 - T + 0.40Zi
2
045
and their slope is 0.40/0.45 = 8/9. The normal will have a slope = —9/8,
and its equation is
X2-0 = ~^{X,-0)
-9
X2 = —5— (see arrow in Figure 16.5)
8
This technique will not tell how far to go in this direction, but we might run
the next experiment with the center of the factorial at ( +1, —9/8). The four
points would be
3^2
1
0 8
0 — 7^
1
2 8
2 — 7^
These points are expressed in terms of the experimental variables, but must
be decoded to see where to set the concentration and position. Using the
coding,
X^ = 4C — 3 or C =
4
X, + 3
X, = 2P - 3 or P = —
2
X, ^2 C P
1 3.1
0 8 4.1
li%
7
0 — 2-
^8 |:1 16
1
2 8 U:1 ll%
2 — 2^ li:l 7
^8 16
16.3 Response-surface experimentation 337
if these settings are possible. Responses may now be taken at these four
points on a new 2^ factorial, and after analysis we can again decide the
direction of steepest ascent. This prodecure continues until the optimum is
obtained.
The above design is sufficient to indicate the direction for subsequent
experiments, but it does not provide a good measure of experimental error
to test the significance of b^, b^, and bj, nor is there any test of how well the
plane approximates the surface. One way to improve on this design is to take
two or more points at the center of the square. By replication at the same
point, an estimate of experimental error is obtained and the average of the
center-point responses will provide an estimate of “goodness of fit” of the
plane. If the experiment is near the maximum response, this center point
might be somewhat above the four surrounding points, which would indicate
the need for a more complex model.
To see how this design would help, consider two observations of response
at the center of the example above. Using the same responses at the vertices
of the square, the results might be those shown in Figure 16.6.
Y ^0 ^2
7.0 1 -1 -1
5.4 1 -1 1
7.1 1 1 -1
6.9 1 1 1
6.6 1 0 0
6.8 1 0 0
1.6
= 0.40
T
- 1.8
-0.45
4
and
SSb, = 6.63(39.8) = 263.87
= 0.40(1.6) = 0.64
SS,^ = -0.45(-1.8) = 0.81
0.81
bi-Fi 1 = TT— = 40.5
^ 0.02
0.32
lack of fit: F2 1 = = 16
0.02
With such a small number of degrees of freedom, only bo shows signifi-
cance at the 5 percent level. However, the tests on the b terms are larger than
the test on lack of fit, which may indicate that the plane
y = 6.63 + 0.40X1 - 0.45X2
is a fair approximation to the surface where this first experiment was run.
Source df SS MS
Total 6 265.98
bo 1 263.87 236.87
bi 1 0.64 0.64
b2 1 0.81 0.81
Residual 3 0.66 —
For this first-order surface, at least four points must be taken to estimate the
four^’s. Since three dimensions are involved, we might consider a 2^ factorial.
As this design has eight experimental conditions at its vertices, the design
often used is a one-half replication of a 2^ factorial with two or more points
at the center of the cube. These six points (two at the center) are sufficient to
estimate all four B’s and to test for lack of fit of this plane to the surface in
three dimensions. After this initial one-half replication of the 2^ the other
half might be run, giving more information for a better fit.
If a plane is not a good fit in two dimensions, a second-order model
might be tried. Such a response surface has the form
Here six B's are to be estimated. The simplest design for this model would be
a pentagon (five points) plus center points. This would yield six or more
responses, and all six S’s could be estimated.
In developing these designs Box and others found that the calculations
can be simplified if the design can be rotated. A rotatable design is one that
has equal predictability in all directions from the center, and the points are
at a constant distance from the center. All first-order designs are rotatable,
as the square and one-half replication of the cube in the cases above. The
simplest second-order design that is rotatable is the pentagon with a point
at the center, as given above.
340 Miscellaneous topics
which has ten unknown coefficients. The cube for a three-dimensional model
has only eight points, so a special design has been developed, called a central
composite design. It is a 2^ factorial with points along each axis at a distance
from the center equal to the distance to each vertex. This gives 8 + 6 = 14
points plus a point at the center makes 15, which is adequate for estimating
the B’s in Equation (16.13). This design is pictured in Figure 16.7.
the shape of this surface can be determined by reducing this equation to what
is called canonical form. This would be
^2 • 1
Figure 16.8
n = 2 3 4 5 6 7 8
0.30 0.35 0.37 0.38 0.39 0.40 0.40
The third cycle is shown in Table 16.11. Here the temperature effect is seen
to be significant, with an increase in temperature giving a higher yield. Once
a significant effect has been found, the first phase of the EVOP procedure has
been completed. The average results at this point are usually displayed on an
EVOP bulletin board as in Figure 16.9.
An EVOP committee usually reviews these data and decides whether
or not to reset the operating conditions. If they change the operating condi-
tions (point 1), then EVOP is reinstated around this new point and the
c
aj
>
0)
Q oo s:
r-
3 V.I
C/5 O
a u,
VH
(N
c (U w (N
cS 3 4-1 3
o c3
00 X C/l
(D
II <1> O
OO 3 tJ) c/5
-*-*
s
bO _o
(N VH O
3
£ w 3 II D +-> a>
~—
o .2 3
<Oi to > dj
ON 44
3 oJ (U to
rt- ^ 3
• • 3 IT)
3 3
OD II II a ;-l
a>
>
jj
3 a>
3
CT3
■H?
CO
. •
fli .. JO
O O 00
bJD
a U
(U
3 3 o
.2 ^ 00
4-<
31 •
> > 3 u u 4H
o' 3
44 X.
a ^ ! 3 oo 00 O O o
^ ^ o^ZZ :z UH PH UH
U-»
*2
»o '3' '3‘ '3‘
o\ ON ON
in O
‘2 *2
•3' ■3- ■3‘ oo
a\ ON ON
o
o
m »n »n l>^
NO NO NO o
m ON ON ON o
2
3 o
o l>r
(/5
4—>
(U
c/5
<D 1 *2
>n »n CO
l>^
+
bO <N 1 rj- ■3' £
c3
4H
TT
03N ON ON
w l>^
a>
2 a
O c/5 > i>? +
l>^
EVOP Work Sheet - First Cycle
U
O 3 3 0-
Sn
<u
bJ)
cd
o
44
3
&
4H
00
4-»
a
O o 00 VH 4-< <D 3
bi) c/5 c/5 3 S
c/5 <1> cb O
_3 3 3 X) > 4H VH 00
to
O 0 a s c« lU
-4—>
1 _o 3
cb 3- c/5 3
c3
Table 16.9
343
c <N
O
_o
’-I-*
00
’>
C/)
-4—^ G 00
<D >r) G
o 'g
00 s:
3
X) G Ui ■>
c/2
03 O
S-i
X> XN
C ttJ (N
03 G G
■*-> C.—
in lo O G
X o <X)
II <1>
(U ^ <D c/2
<N S' bX) ,, bX) -4-4
c II G o
_o B w >^2 G
G .
o
bX)
_G !-(
<u
-<—> G > <^ G 3 > O
c/2 03 JJ (U bX)
UH G
2 (/2 V2 a> 3 G
3 G G II I > G
jj O O <U Oo c/2 G U <3 a> XG
y (1) G o
.W O) (1) 3 '> ’> c ^
G
O
^ -C ^ U g g G W O
S-H
o O
S-i
<U O
O. OH Q £ £ :z: :z; :z UH PX
»r2 •o »o o O sr2
G" Tl- ^ c4 ’-o v-2’ (N
OS OS 0\ I 0\ Os
O
o’
<^2
m
»0 U-2 O *^2 »r2 r-- l>^
m so so’ *0 ■^’ T-4 ITS «r2 G
Os Os Os Os Os
2 «o
73
I
i>r
rs| c/2 o’
D
bX) c/2
-4—>
o3 o 1/2
<L) UH
>^2 irs O *0 <^2 +
■G
c/2
c (U <N Tt rt »r2 o’ 0\ & i>^
>
EVOP Work Sheet—Second Cycle
o Os OS Os I oo OS
W l>^
a C4_l l>^
c/2
<u O o 13"
+
i>^
U P< c/2 G
C o o o o o +
rf so (N o’ »o
-4.J
l>^ +
-l-» Os Os Os I OS Os 2 l>^
2 3 w
3
3
+
l>r
3 U
U
s^|<S
+ O
c/2 bX) l>^ o
G o3 c/2 <u
O g W
C/2 ?£
C/2 <U O G <x>
S ^ o G _o G
TD p & -4—>
G
C 2 2 o3 '—' c/2 (U o (U
o <u O G
3 3 bX)
o S-I
U >2
G
S-I
?S <D
o o G <u -i-<
bX) ^ H c i-i -4-4
G
00 c/2 <U G
_c
G G > S-i <u O
G s-i
O g (U
Table 16.10
:G > >,
344
o <N
c o O
•
o
4-*
d C/5
> 4- >
♦
NO
to
Q
(U
ON
•^
a NO
oo
T3 «o 03
o C/5 5- 1
ccj o
T3 to c 5H
5-(
C O (N
03 ^ c w G
4—> t4H <N
00
M (1>
X 8 O
o3
O
<N
o O
60 ,,
!0
O
60
o3
c/3
4-»
a
c S S c-, C II o
• PH 5-1 O
Tj- (D
Ol O' 4-> G > rA oS
5H
^ 60
o3
> o
c/3 3 o3 60
5H
tr> ty;}" <L) G
4-> 3 G G 13 > 3 o3
o O •O <u c/3 o3 D O 40
a> b’
c/3 3 •^ PH 60 U G G O
b > >
’o’ 03
4.4
U (U O G 5H 5H 5H
5-1 -G o3 ^4 5.4 o3j o a> o o o o
0^ OH OH CLH 0:H PLH UH MH
»o fNj q fN (N
od '=1' d r4
oo ON ON oo ON
HH (N
q m q »n •T)
o
•o’ r--' I_H' NO’ to i-H
ON ON ON 1 oo ON
1 CN »o d
d
I
•o r-- O cn <N r- * IrH
1-H ud NO d r-’ tri >o
ON ON ON 1 oo ON CO
1—H 1 CN
2
13 •o i>^
c/3
<u o l>r
c/3
W) 4->
<u o3 >o •o (N O CC3 o +
VH
<U ON rn cn 1^4
C > Sp.^ oo ON ON OO (ON
<i> o l-H (N w
EVOP Work Sheet—Third Cycle
<4-(
O c/3 o +
>» O
i>r
U 0^ c/3 G
c o o »r) tn q O
•^ + l>^
o d ■^’
•^
4-»
d !0
ON ON ON oo ON JB l>^
+
■r^ CN 3 1^
in jj
3 + 1^
3 i-l|lO
3 U 1^
U
Hc^
+ O
c/3 60 1^ O
c o3 C/3 (U
o 5-1 C/3 &
<L> (/3 o G <u
S 03> C O
G
T3 o JS o3
c 0 <U ‘40 c/3
o <t)
U
o 3 3
(/3
0)
60
o3
<U
JS
o3
V4
(D
4->
a
^ ^ fl> C/5
(U
60 o a 5H
<u 03 G
G
C/3
01 P
C/3
c > 5-1 <u •1.4
o
•^ 03 o 5H
O O O 5-1 c/3 o3 G 60
Table 16.11
o3 .W GH c/3 G
VH
a>
a
o o ^ 3 g o a<u C/3
(U
5H
X
40
O £ Ok Q ^ H OH u
345
346 Miscellaneous topics
94.2 f ♦ 95.7
Pressure 94.8
94.3 4 495.5
Temperature
Figure 16.9
second phase is begun. EVOP is continued again until significant changes are
detected. In fact, it goes on continually, seeking to optimize a process.
This is a very simple example with just two independent variables. The
references should be consulted for more complex situations.
Dp - X p,n
n — 1
Since all these x’s represent the same population, their variances are all alike,
and then
n
n — I
and
16.4 Evolutionary operation (EVOP) 347
The standard deviation of the population can then be written in terms of the
standard deviation of these differences
(jj) can now be estimated from the range of these differences From the
quality control field
A- - J S
estimated by
,
FC/M =
20 20 a}
— ^ (^r) ~ ^
25 25 n
[4 (7Y
Figure 16.10
a\
3 r4
o o^
a
> (/5 1>5
d)
t/D II11
Q gj 3 3 CO
'3
3 0 3
u 3 1
a 3
2n
*3 X
3 *3 II 3^ 3
3
bo o^ 3
3
i->
C/3 3 (S
3 3 •
bO Vi
3
VO 3 B 5^
Vi CO 3
OO O • bO
S
05 to 3
Vi
o C/) c/5 g 3
o o 3 3 3 3 S >
3
o
lo \0
bi) O O
> •3 ^
VO ^ •^
<u c/D o C 3
3 ^ c: ^
<N .a
O' 52 a (N N U 3 3 3 3
CU OH Q
X) c
c o3 X
'3•
Vi
PH z z
03
o o
o O
»o "O
3 CO CO "Tf ON r^, 00
r4
ra 3 Tf Tt co’ o' r^* CO
a> bO
(U u* 3
3 •^
-(.j O
o3 3
o3 Vi TJ
o (U (DCX d) 00 00 O <N 00
a Ov Ov •0 3- d
O s ^ <N
g d) I
*d)
Vi bO
O VH O c/5
EVOP Work Sheet—Block 1 in 2
a O -»-*
C - .4B = i(F4 + ^5 - y2_ - I'aJ = 5^00
A - BC = iC^s + Yi - Y2 - Xs) = 5.40
d)
1/5 o bO c/5
-t-»
d> 3 3 3
U P< Ui Vi
3 p
> ON p p r^. •'N &
< co’ CO in ON '3' w
3 3
O
I—>
-X
u 3 o 3
jj c/5
jj
3 3
s g g c/5
C/5
3 U
U c/5 33 3
3 _oc/5 C/5 l>^
O c/5
3 3^c3 3
bO
TO 0 3^
3
3 g
3
3
22 v^
O c/5
3 3 P 3
c/5 3
>
u O 3O o VH
3
Table 16.12
bO
3 ■> 3 ^
3
CQ<- ♦
g g I 3
3
Vi ,
PH P-i Z Q Z
3
CX ■•- > >.
O
349
VO
m m
C •r5
o
•
on VO
r- CO +1
*rs *
S3 o c/5
(N
‘> -/-< OO OO <N
O
Q
(U II
(N
s
•^
+1
+1
so
3 3
ro ^ O
T3 • o P
CA
I I I
<N '-^3 c/5 Ul
03 o s; c,
T3 o^ s; u <N
C
a
fN ^ 3 (U CN
W
u
-*-» <D P P
bfl CO o3 cp o O
c/5 C o c/5 ID
«4-l
o
II
S'
i-, fO c
_o
It)
bO
o3
c/5
-*-*
s ^ ^ «P
n I I I
_o
-!->
SP >^ IT)
<u
ba
-t—> VH
(U
>
O
O 0 op ^
!/D 03 03 3 o3 (U b£)
3
c/5
D
1/5
P
II s3 sP VH
(U
> 3
p
03 1 I I
o3 O I)
O O <1^ on oo U P
P:
3
U
-r
w <i)
tj)
jrt O (U <U
UH
p
VH
o
;-i ^GG
VH tH ^ o o o ^ ^ op
PH 0^ UH 1^ UH
+ + +
c/5
U op ^
I ^ ^
O• O• 0^• ON« ^• ^ i
O O ^ -H ^
® ; II II II
-P
•W
o op U O
PP ^ ^ oq
p ON VO VO
rj- c/5
os O O »n Tt *0 (N ■/—>
<N o
&
W o
p OO p 00 r-
r4 c4 so cn ON
p
CO _o *
CN
I OO <N
(N o * r4
CJ
3
jj
(N
OO
m
O O
»o »o 3 u
Tj- ON 00 »r5
m *^5
+ 3 *rs I
s *o (N U n +
Tj- Tf- rn o* r^* (N
m’ s
m SO I
(N (N C/D i>^
<U
bO
oS
c/5
-/->
O
+ G GGS
op ^ u
<N VH
(U o o
> <N• <N• <N• o• W
l>^ I , ,
o • l>r i>:r
0) < so rj- m T-H on +
CQ
o ^ 3 2^ o
>> o O
p l>r i>^
^ C o i>r + + +
(U _o •
ID
dj s-* <N
c/:) 3
3 o
i3
3 o GGG P
o3
o
<u
bO I>^
op X X (U
o
o3 3 + + + B
O 3
U
s
B w C/5
c/5
C/5
O
U + + + £
o
oo C ^ 00 330
C s § p
cu o di ^
.2 —
c/5
ig 1^ 103 i|<s
d)
bO
O Cj ' 1 (U
i-i|fS r-ilrs (U
3 3
> T3 o o
bO
03 II II II _§
w C S <u 52
U
O c/5
P
c/5
P
/5 o
s }-i
O
> o op o
^ oq O ^
O O ■§ § s o3 op bO
bO P
'sd C
♦^
>
a>
>
(D S !C + + + o3
o
^ ^ (l> O
VH I-H
-2i
o3
i-i
<L>
PH CU ZQZ op (J
4P
u
3 a > >. -C
H
350
16.5 Analysis of attribute data 351
The first five points (1 through 5) are run as the first block and the last
five points (6 through 10) are run as the second block. Tables 16.12 and 16.13
illustrate this EVOP technique for the second cycle of an example on the
mean surface factor for grinder slivers where the factors are drop-out
temperature A, back-zone temperature B, and atmosphere C. The work
sheets should be self-explanatory as they are based on the design principles
of confounding a 2^ factorial in two blocks of four observations. One could
terminate this process with the second cycle as it shows drop-out temperature
A, back-zone temperature B, and their AB interaction to be significant in
producing changes in the mean surface factor for grinder slivers.
Philosophy
Two assumptions used in the application of the analysis of variance technique
are that (1) the response variable is normally distributed and (2) the variances
of the experimental errors are equal throughout the experiment. In practice
it is often necessary to deal with attribute data where the response variable
is either 0 or 1. In such cases one often records the number of occurrences
of a particular phenomenon or the percentage of such occurrences.
It is well known that the number of occurrences per unit such as defects
per piece, errors per page, customers per unit time, often follow a Poisson
distribution where such response variables are not only nonnormal but
variances and means are equal. When proportions or percentages are used
as the response variable, the data are binomial and again variances are
related to means and the basic assumptions of ANOVA do not hold. Some
studies have shown that lack of normality is not too serious in applying
the ANOVA technique, but most statisticians recommend that a trans-
formation be made on the original data if it is known to be nonnormal.
Recommendations for proper transformations are given in Davies [9].
Another technique, called factorial chi square, has been found very
useful in treating attribute data in industrial problems. This technique is
described by Batson [3], who gives several examples of its application to
actual problems. Although this factorial chi-square technique may not be
as precise as a regular ANOVA on transformed data, its simplicity makes
it well worth consideration for many applied problems.
To illustrate the techniques mentioned above, consider the following
example from a consulting-firm study.
set at two levels giving a 2"^ factorial experiment. Each of the 16 treatment
combinations was run in a completely randomized order, and 20 steel
samples were produced under each of the 16 experimental conditions. The
number of damaged samples in each group of 20 is the recorded variable
in Table 16.14.
Factor A
0 1
Factor B Factor B
Factor C Factor D 0 1 0 1
0 0 0 0 16 20
1 0 0 10 20
1 0 0 0 10 14
1 1 0 12 20
Transformed
Proportion Variable
Treatment P X = arc sin Vp (1) (2) (3) (4) SS
Source df SS MS
A 1 5.12 5.12***
B 1 0.22 0.22*
AB 1 0.35 0.35**
C 1 0.09 0.09
AC 1 0.06 0.06
BC 1 0.03 0.03
D 1 0.03 0.03
AD 1 0.01 0.01
BD 1 0.02 0.02
CD 1 0.02 0.02
Error, pooled 5 0.12 0.02
This analysis shows the highly significant blade size effect, which is
obvious from a glance at Table 16.14. It also shows a significant centering
effect and a blade size-centering interaction that is not as obvious from
Table 16.14.
To apply the factorial chi-square technique to this same example.
Table 16.17 is filled in. The plus and minus signs simply give the proper
weights for the main effects and two-way interactions and T is the value of
the corresponding contrast. D is the total number of samples in the total
experiment; here 20 per cell or 20 x 2"^ = 320 samples. The chi-square (x^)
value is determined by the formula
..2 _
(16.14)
~ (S)(F) ^ ~D
Here N is the total number of items in the whole experiment and S is the
number of occurrences (successes) in the N. F is the number of nonoccur-
rences (or failures). In the example N — 320 samples, 5 = 123 with un-
* * *
(N VO fNj U-) VO Tj- (N VO VO
<S m q VO q
0 T-H OS q
fO 00 d OS d d r4
ON
•n »o ro 00 wo cn •r> 0
9. Ov (N 0 (N <N 0 0 <N r--
<s •ri d d (N d d d d d
Tj”
0 0 0 0 0 0 0 0 0 0
Q <N <N (N (N (N CN tN (N rs| <N
m m m cn ro ro m m CO
Cv Os os WO
Tj- <N 00 r4 00 00 CN
<S VO VO (N
T—H
os m r- ov cn Os WO
<N fS <N
K 1 1 1
+
Os VO 0 00 m r-
1 11 Tl- VO VO rl- VO VO VO w^ WO
1 1 1 1 1 1 1 1 1
tN m •0 VO r4 0 VO OS
+ r- VO r- W-) VO VO VO VO
w
20
+ + + + + + + + + +
0 + + + 1 + + 1 + 1 1
20
+ + 1 + + 1 + 1 + 1
0
20
0 + + 1 1 + 1 1 1 1 +
r4
+ 1 + + 1 + + 1 1 +
0 0
+ 1 + 1 1 + 1 1 + 1
0
'•>6
1—1 0
+ 1 1 + 1 1 + + 1 1
jj
0
3
cn VO
0 + 1 1 1 1 1 1 + + +
H
Ui
0 1 + + + 1 1 1 + + +
+->
(U ’^H
QJ
0 0 1 + + 1 1 1 + + 1 1
GO
Ui T—j 0 1 4- 1 + 1 + 1 1 + 1
O
0
(U 0 0 1 + 1 1 1 + + 1 1 +
d 0
:3
(T T-H
1 1 + + + 1 1 1 1 +
CO
1
0 0 1 1 + 1 + 1 + 1 + 1
u
0
TH 0 1 1 1 + + + 1 + 1 1
0
■t-> 0
0
c3 0
0 1 1 1 1 + + + + + +
UH
(/) 00
-(->
r- 0 °Q C
1-H U C)
,<1^
'O Q to 0
0 -<-> TO ■*—>
ui U 0) T3 13 <0 0
c > d) cd
(U
*T^ -(->
TJ
<D c 3 <u O- ui
d)
U q 0 q q
3 cd C > <U
(U <u D,
<N 3 U co -4-^ q q
05 c
H S U CO t-H
354
16.6 Randomized incomplete blocks—restriction on experimentation 355
, (320)^
= (4.23)(45.75) = 193.52
The 4.23 is constant for these data. The resulting chi-square values are
then compared with with 1 df at the significance level desired. For the
5 percent level, X[i] = 3.84 and again factors A, B, and AB are significant.
As seen in Table 16.17, the calculations for this technique are quite
simple and the results are consistent with those using a more precise method.
Extensions to problems other than 2^ factorials can be found in Batson [3].
Example 16.6 Take the problem of determining the effect on current flow
of four treatments applied to the coils of TV tube filaments. As each treat-
ment application requires some time, it is not possible to run several obser-
vations of these treatments in one day. If days are taken as blocks, all four
treatments must be run in random order on each of several days in order
to have a randomized block design. After checking it is found that even four
treatments cannot be completed in a day; three are the most that can be run.
The question then is: Which treatments are to be run on the first day, which
on the second, and so forth, if information is desired on all four treatments?
The solution to this problem is to use a balanced incomplete block
design. An incomplete block design is simply one in which there are more
treatments than can be put in a single block. A balanced incomplete block
design is an incomplete block design in which every pair of treatments occurs
the same number of times in the experiment. Tables of such designs may be
found in Fisher and Yates [11]. The number of blocks necessary for bal-
ancing will depend upon the number of treatments that can be run in a
single block.
For the example mentioned there are four treatments and only three
treatments can be run in a block. The balanced design for this problem
requires four blocks (days) as shown in Table 16.18.
356 Miscellaneous topics
Treatment
Block
(days) A B C D T,
1 2 — 20 1 29
2 — 32 14 3 49
3 4 13 31 — 48
4 0 23 — 11 34
T, 6 68 65 21 160 = T
In this design only treatments A, C, and D are run on the first day;
B, C, and D on the second day, and so forth. Note that each pair of treat-
ments, such as AB, occurs together twice in the experiment. A and B occur
together on days 3 and 4; C and D occur together on days 1 and 2; and so
on. As in randomized complete block designs, the order in which the three
treatments are run on a given day is completely randomized.
The analysis of such a design is easier if some new notation is intro-
duced. Let
SS total
I j
1^
(160)^
= 3478 - 1344.67
12
16.6 Randomized incomplete blocks—restriction on experimentation 357
y 6)2
cc 1 ^j
(4.8)
^^treatment /c2t
where
Qj = - Z (4.9)
I
Qi = 3(68) - 131 = 73
Q, = 3(65) - 126 = 69
<24 = 3(21) - 112 = -49
0
Note that
Z e. = 0
j=i
Source df SS MS
Totals 11 1344.67
and block degrees of freedom (9) if 4 is subtracted for the four missing values
in the design.
In some incomplete block designs it may be desirable to test for a block
effect. The mean square for blocks was not computed in Table 16.19 since it
had not been adjusted for treatments. In the case of a symmetrical balanced
incomplete randomized block design, where b = t, the block sum of squares
may be adjusted in the same manner as the treatment sums of squares
e; = rT, - I n,jTj
j
Q\ = 3(29) - 92 = -5
Q'2 = 3(49) - 154 = -7
2'3 = 3(48) - 139 = +5
e; - 3(34) - 95 - +7
I e; - 0
(-5)^ + (-7)^ + (5)^ + (7)-
ss block I m- h'/ib - 6.17
i= 1 3(2)4
and
The results of this adjustment and the one in treatments may now be sum-
marized for this symmetrical case as shown in Table 16.20
The terms in parentheses are inserted only to show how the error term
was computed for one adjusted effect and one unadjusted effect
Source df SS MS
Totals 11 1344.67
It should be noted also that the final sum of squares values used in
Table 16.20 to get the mean square values do not add up to the total sum of
squares. This is characteristic of a nonorthogonal design. The F test for
blocks was not run because its value is obviously extremely small, which
indicates no day-to-day effect on current flow.
For nonsymmetrical or unbalanced designs, the general regression
method may be a useful alternative. If contrasts are to be computed for an
incomplete block design, it can be shown that the sum of squares for a
contrast is given by
(4.10)
C, = Q,-Q2 = -166
C2 = QI + Q2- 203 = -158
6^3 = Qi + 62 + 03 — 5^4 = 196
The corresponding sums of squares are
(-166)2
574.08
(2)(3)(2)(4)
(-158)2
173.36
(6)(3)(2)(4)
(196)2
133.39
(12)(24)
880.83
which checks with the adjusted sums of squares for treatments. Comparing
each of the above sums of squares with their 1 df against the error mean
360 Miscellaneous topics
square in Table 16.19 indicates that contrast is the only one of these three
that is significant at the 5 percent level of significance. This indicates a real
difference in current flow between treatments A and B, even though treat-
ments in general showed no significant difference in current flow.
When the conditions for a Latin square are met except for the fact that only
three treatments are possible (for example, because in one block only three
positions are available) and where there are four blocks altogether, the de-
sign is an incomplete Latin square. This design is called a Youden square.
One such Youden square is illustrated in Table 16.21.
Position
Block 1 2 3
I A B C
II D A B
III B C D
IV C D A
Note that the addition of a column (D, C, A, B) would make this a Latin
square if another position were available. A situation calling for a Youden
square might occur if four materials were to be tested on four machines but
there were only three heads on each machine whose orientation might affect
the results.
Example 16.7 The analysis of a Youden square proceeds like the incomplete
block analysis. Assuming hypothetical values for some measured variable
where
^ijk = A Y Pi -\- Tj -\- A- Sijk
with
i=l,...,4 j = 1, 2,..., 4 k = 1, 2, 3
we might have the data of Table 16.22.
In Table 16.22
t = b ^ 4
r = k —
2 = 2
Treatment totals of A, B, C, D are
Tj/.e, 2, -1, -9
16.7 Youden squares 361
Position
Block 1 2 3 T,.
I A = 2 B = 1 C = 0 3
II Z) = -2 4 = 2 B = 2 2
III B = -1 C = -1 D = -3 -5
IV C = 0 D = -4 4 = 2 -2
T.k -1 -2 +1 -2 = T
From this,
SS,„,a, = Z Z Z - “ = 48 - = 47.67
Position effect may first be ignored, since every position occurs once and
only once in each block and once with each treatment, so that positions are
orthogonal to blocks and treatments:
42 1 _ 41
13.67
T “ 3 “ T
For treatment sum of squares adjusted for blocks we get
Q, = 3( + 6) - 3 = 15
= 3( 2) - 0 = 6
Q, = 3(-l) - (-4) = 1
= 3(-9) - (-5) = -22
and
Z e,' = 0
i
SS error = SS....,
'total - SS block
— ‘^‘^treatment
SS (adjusted) SS position
47.67 - 13.67 - 31.08 - 1.17
1.75
The analysis appears in Table 16.23.
362 Miscellaneous topics
Source df SS MS
Totals 11 47.67
as an estimate of error variance, with 5 df. Then the treatment effect is highly
significant. The block mean square is not given, as blocks must be adjusted
by treatments if block effects are to be assessed. The procedure is the same
as shown for incomplete block designs.
PROBLEMS
16.1 One study reports on the breaking strength Y of samples of seven different starch
films where the thickness of each film X was also recorded as a covariate. Results
rounded off considerably are as follows:
Adjusted
Source df df MS
1. Complete this table and set up a test for the effect of starch film or breaking
strength both before and after adjustment for thickness.
2. Write the mathematical model for this experiment, briefly describing each
term.
3. State the assumptions implied in the tests of hypotheses made above and
describe briefly how you would test their validity.
Problems 363
16.2 In studying the effect of two different teaching methods the gain score of each
pupil in a public school test was used as the response variable. The pretest scores
are also available. Run a covariance analysis on the data below to determine
whether or not the new teaching method did show an improvement in average
achievement as measured by this public school test.
Method I Method II
Pretest Pretest
Pupil Score Gain Pupil Score Gain
1 28 5 1 24 7
2 43 -8 2 40 9
3 36 4 3 41 4
4 35 1 4 33 4
5 31 4 5 45 5
6 34 1 6 41 -2
7 33 4 7 41 6
8 38 8 8 33 11
9 39 1 9 41 -1
10 44 -1 10 30 15
11 36 -7 11 45 4
12 21 2 12 50 2
13 34 -1 13 42 9
14 33 1 14 47 3
15 25 7 15 43 -2
16 43 1 16 34 2
17 34 3 17 44 7
18 40 5 18 43 4
19 36 4 19 28 3
20 30 9 20 46 -1
21 35 -6 21 43 2
22 38 0 22 46 -4
23 31 -14 23 21 4
24 41 4
25 35 10
26 27 10
27 45 1
16.3 Data below are for five individuals who have been subjected to four conditions
or treatments represented by the groups or lots 1-4. 7 represents some measure
of an individual supposedly affected by the variations in the treatments of the
four lots. X represents another measure of an individual that may affect the
value of 7 even in the absence of the four treatments. The problem is to deter-
mine whether or not the 7 means of the four groups differ significantly from
each other after the effects of the X variable have been removed.
364 Miscellaneous topics
Groups or Lots
1 2 3 4
X Y X y X Y X y
29 22 15 30 16 12 5 23
20 22 9 32 31 8 25 25
14 20 1 26 26 13 16 28
21 24 6 25 35 25 10 26
6 12 19 37 12 7 24 23
16.4 Consider a two-way classified design with more than one observation per cell.
1. Show the F test for one of the main effects in a covariance analysis if both
factors are fixed.
2. Show the F test for the interaction in a covariance analysis if one factor is
fixed and the other one is random.
16.5 For Example 16.4 in the text, another response variable was measured—the
mean surface factor for heat slivers. Results of the first three cycles at the ten
experimental points gave
Cycle (1) (2) (3) (4) (5) (6) (7) (8) (9) (10)
1 4.0 6.8 5.9 4.5 4.3 4.2 7.8 8.3 11.0 5.9
2 5.2 7.2 6.1 3.7 4.0 4.5 7.0 8.1 12.5 6.3
3 4.8 8.0 7.0 3.6 3.8 4.3 6.5 7.8 12.7 7.0
Set up EVOP work sheets for these data and comment on your results after
three cycles.
16.6 Data on yield in pounds for varying percent concentration at levels 29, 30, and
31 percent, and varying power at levels 400,450, and 500 watts gave the following
results in four cycles at the five points in a 2^ design.
Machine
Lumber
Grade Replication 1 2
A I 8 4
II 4 6
B I 9 3
II 10 4
16.8 Analyze the data of Problem 16.7 using factorial chi square. Compare with the
results of Problem 16.7.
16.9 Data on screen color difference on a television tube measured in degrees Kelvin
are to be compared for four operators. On a given day only three operators can
be used in the experiment. A balanced incomplete block design gave results as
follows:
Operator
Day A B C D
Do a complete analysis of these data and discuss your findings with regard to
differences between operators.
16.10 Run orthogonal contrasts on the operators in Problem 16.9.
16.11 At the Bureau of Standards an experiment was to be run on the wear resistance
of a new experimental shoe leather as compared with the standard leather in use
by the Army. It was decided to equip several men with one shoe of the experi-
mental-type leather and the other shoe of standard leather and after many
weeks in the field compare wear on the two types. Considering each man as a
block, suggest a reasonable number of men to be used in this experiment and
outline its possible analysis.
16.12 If three experimental types of leather were to be tested along with the standard
in Problem 16.11, it is obvious that only two of the four types can be tested on
one man. Set up a balanced incomplete block design for this situation. Insert
some arbitrary numerical values and complete an ANOVA table for your data.
17
Summary and Special Problems
Throughout this book the three phases of an experiment have been empha-
sized: the experiment, the design, and the analysis. At the end of most chap-
ters an outline of the designs considered up to that point has been presented.
It may prove useful to the reader to have a complete outline as a summary of
the designs presented in Chapters 1-15.
This is not the only way such an outline could be constructed, but it
represents an attempt to see each design as part of the overall picture of
designed experiments. A look at the outline will reveal that the same design
is often used for an entirely different experiment, which simply illustrates
the fact that much care must be taken to spell out the experiment and its
design or method of randomization before an experiment is performed.
For each experiment, the chapter reference in this book is given.
After the summary several special problems are presented. Many do
not fit any one pattern or design but may be of interest as they represent
the kinds of problems actually encountered in practice. These problems
have no “pat” solutions and the way they are set up may depend a great
deal upon assumptions the reader may make. They should serve to evoke
some good discussion about designing an experiment to solve a given prac-
tical problem.
366
X X
<u
0
m 1)- ■G- X X o^
a> Vi Vi Vi Vi Vi Vi Vi Vi Vi
G
(*V <l-> d) d) d) d) d) d) d) d) d)
;-H ->-> ■1—> -11 -1—1 -1-1 -1-1 1-1 1-1
03 CX GH Cv Pi a PH Pi Pi Pi
JZ C-I-M G G G G cS G G a G
D
u X X X X X X X X X
U u U U U U U U U
TD
O
175
P _i • •
+-<
d)
<NI
X)
0 G
Vi
d)
• -
(U 0 G 1-H <N
G (U -,—1 cq
< Vi
C GO • • rs 0
< > d>
1-1
> _C
Vi • <I
C/D <
< •§ 0
< < 0
x' > 0 +
0 > ^ Z > > X! G
C/D >
0
Z
<
0
^
^
GO
< o o X
1-1
X
0
'3
x' Z
0 o'
(N
cq
X
0
X 03 :z 1-1 0
z d) 0"
a < o3
< < <
II G
Vi
< < o3
:>
g
♦—(
> 03
>1 :>
0
1
cG
0
Vi
^
03
03 'o
03 >
0
r-
C/D
a> 0
^^ D 0
cq +-<
_G
and
Vi A\ ^ X
1 cii X a ^ Z o3 Z 03
CO
(D
-G
d)
CO
Vi
(D H iX 0 H 00 3 < < G 0 G c2
C
. o
0 d X 0 d X lx d X cJ
(N m
CO
+
s CO
d»
Ui
3 +
CO
o3
+ D +
C cr cq
TD
d> co“ TDg c/5
C
d)
CO
N a
u< + d)
'T3
+ +
Q
I + g d)
a
d)
00
+
•“-5 G
O
>H
^
^ +
(T '
KT
o cq
T3
+ 2
G
i +3 +
^
•-» - -v
d) dT (U
C/D
G H“ P
Vi ^
G
d) ^ O
>>'"^03 CCi, ^ ^ d>
'a a
^ (U
S, CIH HJG + + G
o + .a G 3
+
G, 13 Vi
d) 03
9- S o CT
a 4—» -1 d) d)
CO CO
'a. ^ o ' I 8 CJ
CO
3.
a o oo I
(U
Oi II
■X Vi
O
G d>
G
0
G
0
§ a d)
aO H--^ "Sc3 II o £ C
U £ G
Vi
s ao -
S o "-I
<N
X
m
CNI m
CO "O
Vi d)
O
1—>
CO
(75
o O
Vi
C ,G
o
d)
1>
avC Vi
o n
^
^
Cd
d)
'll
o
V11
a s
w
X O
n 03
GO
_C
O tx
H <
367
(N 'Cj- lO
■r—1 ▼-H
^ 8 ;-! VH
'—1
"XX o X _o _3 X
C/D (L>
> z cd o X3 _3
—H
3 C/D
•
c/5 O
c< < CD o O3 o3 'CH _o
w
-*-H
3o 3o
H—<
'CH X3
C/D
c/3
■(—I o s Xo
1
o AJ
o CQ
1-^^
13 < a
o
o3
H—<
X
3 (L> o3 VH H—<
X
X
3O (L>
VH
o3H—>
_3 <4—»
c/5
c CO ■
"d
X G OQ 3 X CJ _3 X "3
< Ui cO
O CJ OH
CO
<N m
8
cO
<N
JS-
CQ
CO CO oT
c/5
+ C< + <D
(/5
00
S-I
•“-5
+ •it
+ •2
CQ
cd
3
cq -it C<
cq •it lo CT
'' 0 cq c/3
CQ
+
X
0
X
’OH +
+
c +
^ X
X ^ c/5
_o 3
+ '3'
ccT bX) _o + X X 3
H—> of'
_3 XJ 0
cq § cO
o
3 + X
"OH
-4—)
•■-«3
0 CQ-
3 +
c + +
CJD
c/3
3
3 +
X c<
<N
S ttf , O .-.
CC!, (U ^
V-H
3 'OH
c/5 1. + ^
Q
+
.2
3 cZ
+ _G ^
.2 +
I
cO +
X
c/5 8 cs, C “-H
3 x: 0 4-<
CQ O
’o. =t 3, (N m
_o
_o 3 3 + (D
3 _o 2 II ^ II 4-< C G
X 3" X •*«4» + H-H w =■ <u C-)
~o 3 + -HH
3
3 o > cO
H—» ccj <U
3
N
3 =t 'OH
_3
3
II
-it
l-H
lU
CO o (JL| I>-i
cO X)
X 3
S II
a
C rT m
e 0 S
0 qx! 3
0
X 3
3 u X“ 1—1
3
X cd X
<N
c
a>
s
• T*H
!-i
<U
a
X
W
368
Chapter 12
7l- 1 1
DC T-H 'T—H
VH C/5 -r-H DC
l-l
03T—H OJ ^ !-i
CD
H—* <l> ^ - 03
-(—>
•I—I
73
OH OH O ^ c OH
^ O o 7:3
o3 o3 3 5 03 o3 !=1
HO o3 HO aq o3
U u U U U U
replications and positions
< < g
a. Factorial ANOVA with
> > o
'S DC
<
> o o
> o o <
^
<
o
OH
< Ln
:z
"d
< > > O
<
VH *rl ^
> o O a. O
H—‘ DC o ^C/5
o 'iO
—>
CD
M-H CD
< Ci? < g
< 7D c/o TD
<U
crc
JD
0
73031 3
^ 8
O DC O 1 -1—> HO
T:5 DC ■»—> Vc ^
CD Q-) O CD o (D 03 <1^ .Id
C/5 on
<V (D
z c3 o3 3 3
3.
(N m <N m
-ii
q*;s:
oi"
+ CO
-iji s +
^ I
+ Bj
CO
+ S
T3 + £ r ■“
CD ^ 03 — u
N —
_N —
+ CO
+ + +^
I cq" ^ +
C<
a ^ + .i > +
o
c +•- -O
^ C< +
'§ +
c
o U
+ a ^ u +^ + o
0-) 03 3, +
. Latin square
^ ^ X! 03
a. Complete
(U (D 3. 73
>> + cq 03 H—>3, +’ ^
;-l H—1
13 3, o3 <U N 03 c3 03 II
+
Yijkm =
ID + .^
DH II 3 'DH II
;S ^ a cr OH B 3, 'OH
a 'L C/i R ^
03
’HH II o a £
•“-i
c/5 a £
^ II -g rl O 3 73 o o
s ^ U c ■§ U >< ao --s.s a u U
O ^03 3 o3 o3
U Cti d 1—1 d u d O d
m m (N rd
03
O
o a
VH
Ut
UH o
<u H—<
CD
CHH
'O 7D
<u
-(—> 03
on
CD 03
z
pp u
369
370 Summary amd special problems
SPECIAL PROBLEMS
17.1 Three different time periods are chosen in which to plant a certain vegetable:
early May, late May, and early June. Plots are selected in early May and five
different fertilizers are randomly tried in each plot. This is repeated in late May
and again in early June. The criterion is the number of bushels per acre at harvest
time. Assuming r plots for each fertilizer can be used at each planting:
a. Discuss the nature of this experiment.
b. Recommend a value for r to make this an acceptable study. (Time periods
and fertilizers are fixed.)
17.2 In the study of the factors affecting the surface finish of steel in Problem 15.19
there was a sixth factor—factor A (the flux type)—also set at two levels. It was
possible to run all 64 (2^) treatment combinations except for the fact that the
experimenter wanted a complete replication of the experiment that was to be
run. Thus, it was decided to do a one-half replicate of the 2^ and then replicate
the same block at another time. It was also learned that first a flux type had to
be chosen, and then a slab width, and then the other four factors randomized.
After this the experimenter would run the other slab width and then repeat the
experiment with the other flux type.
Outline a data format for this problem and show what its ANOVA table
might look like.
17.3 In Problem 15.18 on the 81 engines it took about a year to gather all the
data. When the data were ready they were found to be incomplete at the low-
temperature level so the experiment was really a one-third replication of a 3“^ x 2
factorial instead of the 3^ factorial. How would you analyze the data if one
whole level in your answer to this problem were missing?
17.4 In a food industry there was a strong feeling that viscosity readings on a slurry
were not consistent. Some claimed that different laboratories gave different
readings. Others said it was the fault of the instruments in these laboratories.
And still others claimed it was the fault of the operators who read the instruments.
To determine where the main sources of variability were, slurries from two prod-
ucts with vastly different viscosities were taken. Each sample of slurry was
mixed well and then divided into four jars, one of which was sent to each of
four laboratories for analysis. In each laboratory the jar was divided into eight
samples in order for each of two randomly chosen operators within that labora-
tory to test two samples on each of two randomly chosen test instruments.
Considering the laboratories as fixed, the two products as fixed, operators as
random, and instruments as random, show a data layout for this problem. Show
its model and its ANOVA outline and discuss any concerns about the tests to
be made.
17.5 A total of 256 seventh grade students with equal numbers of boys and girls and
representing both high and low socioeconomic status (SES) was subjected to a
mathematics lesson that was believed to be either interesting or uninteresting
under one of four reinforcement conditions: no knowledge of results, immediate
verbal knowledge of results, delayed verbal knowledge of results, and delayed
written knowledge of results. Task interest and type of reinforcement emerged
Summary and special problems 371
17.6 In testing for a significant effect of treatments on some measured variable Y the
resulting F test was almost, but not quite, significant at the 5 percent level (say,
F = 4.00 with 3 and 8 df). Now the experimenter repeats the whole experiment
on another set of independent observations under the same treatments and again
the F fails to reach the 5 percent level (say, F = 4.01 with 3 and 8 df). What
conclusion—if any—would you reach with regard to the hypothesis of equal
treatment effects? Briefly explain your reasoning.
17.7 Six different formulas or mixes of concrete are to be purchased from five competing
suppliers. Tests of crushing strength are to be made on two blocks constructed
according to each formula-supplier combination. One block will be poured,
hardened, and tested for each combination before the second block is formed,
thus making two replications of the whole experiment. Assuming mixes and
suppliers fixed and replications random, set up a mathematical model for this
situation, outline its ANOVA table, and show what F tests are appropriate.
17.8 Four chemical treatments A, B, C, and D are prepared in three concentrations
(3 to 1, 2 to 1, and 1 to 1) and two samples of each combination are poured; then
three determinations of percent of pH are made from each sample. Considering
treatments and concentrations fixed and samples and determinations random,
set up the appropriate ANOVA table and comment on statistical tests to be
made. Also explain in some detail how you would proceed to set 95 percent
confidence limits on the average percent of pH for treatment B, concentration
2 to 1.
17.9 A graduate student carried out an experiment involving six experimental con-
ditions. From an available group of subjects she assigned five men and five
women at random to each of the six conditions, using 60 subjects in all. Upon
completion of the experiment, she carried out the following analysis to test the
differences among the condition means:
Source df
Between conditions 5
F = —^ with 5 and 54 df
Within conditions 54 MS^
Total 59
372 Summary and special problems
Ai
Br B2 B, B2
D, 2 2 1 5
Cl
D2 2 1 2 3
D, 4 2 5 1
Co
D2 1 1 2 2
Glossary of Terms
373
374 Glossary of terms
Mean. Of a sample:
Y = t YJn
i— 1
Of a population:
fi = E{Y).
Mean square. An unbiased estimate of a population variance. Determined
by dividing a sum of squares by its degree of freedom.
Minimum variance estimate. If two or more estimates u^, Uj,... are
made of the same parameter 6, the estimate with the smallest variance is
called the minimum variance estimate.
Mixed model. The model of a factorial experiment in which one or more
factors are at fixed levels and at least one factor is at random levels.
Nested experiment. An experiment in which the levels of one factor are
chosen within the levels of another factor.
Nested-factorial experiment. An experiment in which some factors are
factorial or crossed with others and some factors are nested within others.
Operating characteristic curve. A plot of the probability of acceptance of a
hypothesis under test versus the true parameter of the population.
Orthogonal contrasts. Two contrasts are said to be orthogonal if the prod-
ucts of their corresponding coefficients add to zero.
Orthogonal polynomials. Sets of polynomials used in regression analysis
such that each polynomial in the set is orthogonal to all other polynomials
in the set.
Parameter. A characteristic of a population, such as the population mean
or variance.
Point estimate. A single statistic used to estimate a parameter.
Power of a test. A plot of the probability of rejection of a hypothesis under
test versus the true parameter. It is the complement of the operating char-
acteristic curve.
The block in a confounded design that contains the treat-
Principal block.
ment combination in which all factors are at their lowest level.
Random model. A model of a factorial experiment in which the levels of
the factor are chosen at random.
Random sample. A sample in which each member of the population
sampled has an equal chance of being selected in this sample.
Randomized block design. An experimental design in which the treatment
combinations are randomized within a block and several blocks are run.
Repeated-measures design. A design in which measurements are taken on
the experimental units more than once.
376 Glossary of terms
SS = t (Fi - Yf
i= 1
2 T,UAY,-Yf SS
f! — 1 df
Of a population:
= E(Y - nf.
Youden square. An incomplete Latin square.
References
377
378 References
17. McCall, C. H., Jr., “Linear Contrasts, Parts I, II, and III,” Industrial Quality Control
(July-Sept., 1960).
18. Miller, L. D., “An Investigation of the Machinability of Malleable Iron Using
Ceramic Tools,” unpublished MSIE thesis, Purdue University (1959).
19. Ostle, B., Statistics in Research (2nd Ed.). Ames, Iowa: Iowa State University Press
(1963).
20. Owen, D. B., Handbook of Statistical Tables. Boston: Addison-Wesley Publishing
Company, Inc. (1962).
21. Scheffe, H., “A Method for Judging All Contrasts in the Analysis of Variance,”
Biometrics, vol. XL (June, 1953).
22. Snedecor, G. W., and W. C. Cochran, Statistical Methods (6th Ed.). Ames, Iowa:
Iowa State University Press (1967).
23. Tanur, Judith M., editor, and others. Statistics; A Guide to the Unknown (2nd Ed.)
San Francisco: Holden-Day (1978).
24. Winer, B. J., Statistical Principles in Experimental Design (2nd Ed.). New York:
McGraw-Hill Book Company (1971).
25. Wortham, A. W., and T. E. Smith, Practical Statistics in Experimental Design.
Dallas: Dallas Publishing House (1960).
26. Yates, F., Design and Analysis of E act or ial Experiments. London: Imperial Bureau
of Soil Sciences (1937).
‘ . -■' f. 1^ I iT T. W,- - ■• ./i.»> J. Vitt ■>■» r .4 .'-ij. *■ ' :'V • . ■ ■
.s # ;■
•’ 'J , -V*^^
'.T^-v
•
L.^-'"f-
EG -IpIP' ifT r.
/ • • " • . *' f, **>' J't
v*i* ,'•. ^> ]
fee. ■
L>i'*'/,'’■ ; ;A,. '■ ,iitf'
e .i ':i:-A ^
•TV-' AvyS*"**,
14'^-
rjH.Vi.'r
t'e- •
■■■
'
^ A r.’^'-j'.i'i
' '} 1: ,^.r v. /i.^;.'yv:.''.5 .,';'^^^^l' 'y
r-' .. ‘
5' •. ''tf-V'.
Statistical Tables
Table A Areas under the Normal Curve* (Proportion of Total Area under the
Curve from — oo to Designated Z Value)
z 0.09 0.08 0.07 0.06 0.05 0.04 0.03 0.02 0.01 0.00
-3.5 0.00017 0.00017 0.00018 0.00019 0.00019 0.00020 0.00021 0.00022 0.00022 0.00023
-3.4 0.00024 0.00025 0.00026 0.00027 0.00028 0.00029 0.00030 0.00031 0.00033 0.00034
-3.3 0.00035 0.00036 0.00038 0.00039 0.00040 0.00042 0.00043 0.00045 0.00047 0.00048
-3.2 0.00050 0.00052 0.00054 0.00056 0.00058 0.00060 0.00062 0.00064 0.00066 0.00069
-3.1 0.00071 0.00074 0.00076 0.00079 0.00082 0.00085 0.00087 0.00090 0.00094 0.00097
-3.0 0.00100 0.00104 0.00107 0.00111 0.00114 0.00118 0.00122 0.00126 0.00131 0.00135
-2.9 0.0014 0.0014 0.0015 0.0015 0.0016 0.0016 0.0017 0.0017 0.0018 0.0019
-2.8 0.0019 0.0020 0.0021 0.0021 0.0022 0.0023 0.0023 0.0024 0.0025 0.0026
-2.7 0.0026 0.0027 0.0028 0.0029 0.0030 0.0031 0.0032 0.0033 0.0034 0.0035
-2.6 0.0036 0.0037 0.0038 0.0039 0.0040 0.0041 0.0043 0.0044 0.0045 0.0047
-2.5 0.0048 0.0049 0.0051 0.0052 0.0054 0.0055 0.0057 0.0059 0,0060 0.0062
-2.4 0.0064 0.0066 0.0068 0.0069 0.0071 0.0073 0.0075 0.0078 0.0080 0.0082
-2.3 0.0084 0.0087 0.0089 0.0091 0.0094 0.0096 0.0099 0.0102 0.0104 0.0107
-2.2 0.0110 0.0113 0.0116 0.0119 0.0122 0.0125 0.0129 0.0132 0.0136 0.0139
-2.1 0.0143 0.0146 0.0150 0.0154 0.0158 0,0162 0.0166 0.0170 0.0174 0.0179
-2.0 0.0183 0.0188 0.0192 0.0197 0.0202 0.0207 0.0212 0.0217 0.0222 0.0228
-1.9 0.0233 0.0239 0.0244 0.0250 0.0256 0.0262 0.0268 0.0274 0.0281 0.0287
-1.8 0.0294 0.0301 0.0307 0,0314 0.0322 0.0329 0.0336 0.0344 0.0351 0.0359
-1.7 0.0367 0.0375 0.0384 0.0392 0.0401 0.0409 0.0418 0.0427 0.0436 0.0446
-1.6 0.0455 0.0465 0.0475 0,0485 0.0495 0.0505 0.0516 0.0526 0.0537 0.0548
-1.5 0.0559 0.0571 0.0582 0.0594 0,0606 0.0618 0.0630 0.0643 0.0655 0.0668
-1.4 0.0681 0.0694 0.0708 0.0721 0.0735 0.0749 0.0764 0.0778 , 0.0793 0.0808
-1.3 0.0823 0.0838 0.0853 0.0869 0.0885 0.0901 0.0918 0.0934 0.0951 0.0968
-1.2 0.0985 0.1003 0.1020 0.1038 0.1057 0.1075 0.1093 0.1112 0.1131 0.1151
-1.1 0.1170 0.1190 0.1210 0.1230 0.1251 0 1271 0.1292 0.1314 0.1335 0.1357
-1.0 0.1379 0.1401 0.1423 0.1446 0.1469 0.1492 0.1515 0.1539 0,1562 0.1587
-0.9 0.1611 0.1635 0.1660 0.1685 0.1711 0.1736 0.1762 0.1788 0.1814 0.1841
-0.8 0.1867 0.1894 0.1922 0.1949 0.1977 0.2005 0.2033 0.2061 0.2090 0.2119
—0.7 0.2148 0.2177 0.2207 0.2236 0.2266 0.2297 0.2327 0.2358 0.2389 0.2420
-0.6 0.2451 0.2483 0.2514 0.2546 0.2578 0.2611 0.2643 0.2676 0.2709 0.2743
-0.5 0.2776 0.2810 0.2843 0.2877 0.2912 0.2946 0.2981 0.3015 0.3050 0.3085
-0.4 0.3121 0.3156 0.3192 0.3228 0.3264 0.3300 0.3336 0.3372 0.3409 0.3446
-0.3 0.3483 0.3520 0.3557 0.3594 0.3632 0.3669 0.3707 0.3745 0.3783 0.3821
-0.2 0.3859 0.3897 0.3936 0.3974 0.4013 0.4052 0.4090 0.4129 0.4168 0,4207
-0.1 0.4247 0.4286 0.4325 0.4364 0.4404 0.4443 0.4483 0.4522 0.4562 0.4602
-0.0 0.4641 0.4681 0.4721 0.4761 0.4801 0.4840 0.4880 0.4920 0.4960 0.5000
* Adapted from E. L. Grant, Statistical Quality Control (2nd Ed.). New York: McGraw-
Hill Book Company, Inc. (1952), Table A, pp. 510-511. Reproduced by permission of the
publisher.
380
Statistical tables 381
Table A {Continued)
z 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0,08 0.09
+0.0 0.5000 0.6040 0.5080 0.5120 0.5160 0.5199 0.5239 0.5279 0.5319 0.5359
+0.1 0.6398 0.5438 0.5478 0.5517 0.5557 0.5596 0.5636 0.5675 0.5714 0.5753
+0.2 0.5793 0.5832 0.5871 0.5910 0.5948 0.5987 0.6026 0.6064 0.6103 0.6141
+0.3 0.6179 0.6217 0.6255 0.6293 0.6331 0.6368 0.6406 0.6443 0.6480 0.6517
+0.4 0.6554 0.6591 0.6628 0.6664 0.6700 0.6736 0.6772 0.6808 0.6844 0.6879
+0.5 0.6915 0.6950 0.6985 0.7019 0.7054 0.7088 0.7123 0.7157 0.7190 0.7224
+0.6 0.7257 0.7291 0.7324 0.7357 0.7389 0.7422 0.7454 0.7486 0.7517 0.7549
+0.7 0.7580 0.7611 0.7642 0.7673 0.7704 0.7734 0.7764 0.7794 0.7823 0.7852
+0.8 0.7881 0.7910 0.7939 0.7967 0.7995 0.8023 0.8051 0.8079 0.8106 0.8133
+0.9 0.8159 0.8186 0.8212 0.8238 0.8264 0.8289 0.8315 0.8340 0.8365 0.8389
+ 1.0 0.8413 0.8438 0.8461 0.8485 0.8508 0.8531 0.8554 0.8577 0,8599 0.8621
+ 1.1 0.8643 0.8665 0.8686 0.8708 0.8729 0.8749 0.8770 0.8790 0.8810 0.8830,
+ 1.2 0.8849 0.8869 0.8888 0.8907 0.8925 0.8944 0.8962 0.8980 0.8997 0.9015
+ 1.3 0.9032 0.9049 0.9066 0.9082 0.9099 0.9115 0.9131 0.9147 0.9162 0.9177
+ 1.4 0.9192 0.9207 0.9222 0.9236 0.9251 0.9265 0.9279 0.9292 0.9306 0.9319
+ 1.5 0.9332 0.9345 0.9357 0.9370 0.9382 0.9394 0.9406 0.9418 0.9429 0.9441
+ 1.6 0.9452 0.9463 0.9474 0.9484 0.9495 0.9505 0.9515 0.9525 0.9535 0.9545
+ 1.7 0.9554 0.9564 0.9573 0.9582 0.9591 0.9599 0.9608 0.9616 0.9625 0.9633
+ 1.8 0.9641 0.9649 0.9656 0.9664 0.9671 0.9678 0.^68^ 0.9693 0.9699 0.9706
+ 1.9 0.9713 0.9719 0.9726 0.9732 0.9738 0.9744( 0.9750 0.9756 0.9761 0.9767
+2.0 0.9773 0.9778 0.9783 0.9788 0.9793 0.9798^ 0r0803 0.9808 0.9812 0.9817
+2.1 0.9821 0.9826 0.9830 0.9834 0.9838 0.9842 0.9846 0.9850 0.9854 0.9857
+2.2 0.9861 0.9864 0.9868 0.9871 0.9875 0.9878 0.9881 0.9884 0.9887 0.9890
+2.3 0.9893 0.9896 0.9898 0.9901 0.9904 0.9906 0.9909 0.9911 0.9913 0.9916
+2.4 0.9918 0.9920 0.9922 0.9925 0.9927 0.9929 0.9931 0.9932 0.9934 0.9936
+2.6 0.9938 0.9940 0.9941 0.9943 0.9945 0.9946 0.9948 0.9949 0.9951 0.9952
+2.6 0.9953 0.9955 0.9956 0.9957 0.9959 0.9960 0.9961 0.9962 0.9963 0.9964
+2.7 0.9965 0.9966 0.9967 0.9968 0.9969 0.9970 0.9971 0.9972 0.9973 0.9974
+2.8 0.9974 0.9975 0.9976 0.9977 0.9977 0.9978 0.9979 0.9979 0.9980 0.9981
+2.9 0.9981 0.9982 0.9983 0.9983 0.9984 0.9984 0.9985 0.9985 0.9986 0.9986
+3.0 0.99865 0.99869 0.99874 0.99878 0.99882 0.99886 0.99889 0.99893 0.99896 0.99900
+3.1 0.99903 0.99906 0.99910 0.99913 0.99915 0.99918 0.99921 0.99924 0.99926 0.99929
+3.2 0.99931 0.99934 0.99936 0.99938 0.99940 0.99942 0.99944 0.99946 0.99948 0.99950
+3.3 0.99952 0.99953 0.99955 0.99957 0.99958 0.99960 0.99961 0.99962 0.99964 0.99965
+3.4 0.99966 0.99967 0.99969 0.99970 0.99971 0.99972 0.99973 0.99974 0.99975 0.99976
+3.5 0.99977 0.99978 0.99978 0.99979 0.99980 0.99981 0.99981 0.99982 0.99983 0.99983
B Student’s t Distribution
Percentile point
70 80 90 95 97.5 99 99.5
382
Table C Chi Square*
Probability
V
0.99 0.98 0.95 0.90 0.80 0.20 0.10 0.05 0.02 0.01 0.001
1 0.0n57 0.0^628 0.00393 0.0158 0.0642 1.642 2.706 3.841 5.412 6.635 10.827
2 0.0201 0.0404 0.103 0.211 0.446 3.219 4.605 5.991 7.824 9.210 13.815
3 0.115 0.185 0.352 0.584 1.005 4.642 6.251 7.815 9.837 11.341 16.268
4 0.297 0.429 0.711 1.064 1.649 5.989 7.779 9.488 11.668 13.277 18.465
5 0.554 0.752 1.145 1.610 2.343 7.289 9.236 11.070 13.388 15.086 20.517
6 0.872 1.134 1.635 2.204 3.070 8.558 10.645 12.592 15.033 16.812 22.457
7 1.239 1.564 2.167 2.833 3.822 9.803 12.017 14.067 16.622 18.475 24.322
8 1.646 2.032 2.733 3.490 4.594 11.030 13.362 15.507 18.168 20.090 26.125
9 2.088 2.532 3.325 4.168 5.380 12.242 14.684 16.919 19.679 21.666 27.877
10 2.558 3.059 3.940 4.865 6.179 13.442 15.987 18.307 21.161 23.209 29.588
11 3.053 3.609 4.575 5.578 6.989 14.631 17.275 19.675 22.618 24.725 31.264
12 3.571 4.178 5.226 6.304 7.807 15.812 18.549 21.026 24.054 26.217 32.909
13 4.107 4.765 5.892 7.042 8.634 16.985 19.812 22.362 25.472 27.688 34.528
14 4.660 5.368 6.571 7.790 9.467 18.151 21.064 23.685 26.873 29.141 36.123
15 5.229 5.985 7.261 8.547 10.307 19.311 22.307 24.996 28.259 30.578 37.697
16 5.812 6.614 7.962 9.312 11.152 20.465 23.542 26.296 29.633 32.000 39.252
17 6.408 7.255 8.672 10.085 12.002 21.615 24.769 27.587 30.995 33.409 40.790
18 7.015 7.906 9.390 10.865 12.857 22.760 25.989 28.869 32.346 34.805 42.312
19 7.633 8.567 10.117 11.651 13.716 23.900 27.204 30.144 33.687 36.191 43.820
20 8.260 9.237 10.851 12.443 14.578 25.038 28.412 31.410 35.020 37.566 45.315
21 8.897 9.915 11.591 13.240 15.445 26.171 29.615 32.671 36.343 38.932 46.797
22 9.542 10.600 12.338 14.041 16.314 27.301 30.813 33.924 37.659 40.289 48.268
23 10.196 11.293 13.091 14.848 17.187 28.429 32.007 35.172 38.968 41.638 49.728
24 10.856 11.992 13.848 15.659 18.062 29.553 33.196 36.415 40.270 42.980 51.179
25 11.524 12.697 14.611 16.473 18.940 30.675 34.382 37.652 41.566 44.314 52.620
26 12.198 13.409 15.379 17.292 19.820 31.795 35.563 38.885 42.856 45.642 54.052
27 12.879 14.125 16.151 18.114 20.703 32.912 36.741 40.113 44.140 46.963 55.476
28 13.565 14.847 16.928 18.939 21.588 34.027 37.916 41.337 45.419 48.278 56.893
29 14.256 15.574 17.708 19.768 22.475 35.139 39.087 42.557 46.693 49.588 58.302
30 14.953 16.306 18.493 20.599 23.364 36.250 40.256 43.773 47.962 50.892 59.703
* Abridged from R. A. Fisher and F. Yates, Statistical Tables for Biological, Agricultural and
Medical Research. Edinburgh and London: Oliver & Boyd, Ltd (1953), Table IV. Reproduced
by perm.ission of the authors and publishers. For larger values of v, the expression —
■sj2v — 1 may be used as a normal deviate with unit variance, remembering that the probability
for corresponds with that of a single tail of a normal curve.
383
istribution*
0 F
df for numerator
1 2 3 4 5 6 7 8 9 10 11 12
5.83 7.50 8.20 8.58 8.82 8.98 9.10 9.19 9.26 9.32 9.36 9.41
1 39.9 49.5 53.6 55.8 57.2 58.2 58.9 59.4 59.9 60.2 60.5 60.7
161 200 216 225 230 234 237 239 241 242 243 244
2.57 3.00 3.15 3.23 3.28 3.31 3.34 3.35 3.37 3.38 3.39 3.39
2 8.53 9.00 9.16 9.24 9.29 9.33 9.35 9.37 9.38 9.39 9.40 9.41
18.5 19.0 19.2 19.2 19.3 19.3 19.4 19.4 19.4 19.4 19.4 19.4
98.5 99.0 99.2 99.2 99.3 99.3 99.4 99.4 99.4 99.4 99.4 99.4
2.02 2.28 2.36 2.39 2.41 2.42 2.43 2.44 2.44 2.44 2.45 2.45
3 5.54 5.46 5.39 5.34 5.31 5.28 5.27 5.25 5.24 5.23 5.22 5.22
10.1 9.55 9.28 9.12 9.10 8.94 8.89 8.85 8.81 8.79 8.76 8.74
34.1 30.8 29.5 28.7 28.2 27.9 27.7 27.5 27.3 27.2 27.1 27.1
1.81 2.00 2.05 2.06 2.07 2.08 2.08 2.08 2.08 2.08 2.08 2.08
4 4.54 4.32 4.19 4.11 4.05 4.01 3.98 3.95 3.94 3.92 3.91 3.90
7.71 6.94 6.59 6.39 6.26 6.16 6.09 6.04 6.00 5.96 5.94 5.91
21.2 18.0 16.7 16.0 15.5 15.2 15.0 14.8 14.7 14.5 14.4 14.4
1.69 1.85 1.88 1.89 1.89 1.89 1.89 1.89 1.89 1.89 1.89 1.89
5 4.06 3.78 3.62 3.52 3.45 3.40 3.37 3.34 3.32 3.30 3.28 3.27
6.61 5.79 5.41 5.19 5.05 4.95 4.88 4.82 4.77 4.74 4.71 4.68
16.3 13.3 12.1 11.4 11.0 10.7 10.5 10.3 10.2 10.1 9.96 9.89
1.62 1.76 1.78 1.79 1.79 1.78 1.78 1.77 1.77 1.77 1.77 1.77
6 3.78 3.46 3.29 3.18 3.11 3.05 3.01 2.98 2.96 2.94 2.92 2.9C
5.99 5.14 4.76 4.53 4.39 4.28 4.21 4.15 4.10 4.06 4.03 4.00
13.7 10.9 9.78 9.15 8.75 8.47 8.26 8.10 7.98 7.87 7.79 7.72
1.57 1.70 1.72 1.72 1.71 1.71 1.70 1.70 1.69 1.69 1.69 1.68
7 3.59 3.26 3.07 2.96 2.88 2.83 2.78 2.75 2.72 2.70 2.68 2.67
5.59 4.74 4.35 4.12 3.97 3.87 3.79 3.73 3.68 3.64 3.60 3.57
12.2 9.55 8.45 7.85 7.46 7.19 6.99 6.84 6.72 6.62 6.54 6.47
1.54 1.66 1.67 1.66 1.66 1.65 1.64 1.64 1.64 1.63 1.63 1.62
8 3.46 3.11 2.92 2.81 2.73 2.67 2.62 2.59 2.56 2.54 2.52 2.50
5.32 4.46 4.07 3.84 3.69 3.58 3.50 3.44 3.39 3.35 3.31 3.28
11.3 8.65 7.59 7.01 6.63 6.37 6.18 6.03 5.91 5.81 5.73 5.67
1.51 1.62 1.63 1.63 1.62 1.61 1.60 1.60 1.59 1.59 1.58 1.58
9 3.36 3.01 2.81 2.69 2.61 2.55 2.51 2.47 2.44 2.42 2.40 2.38
5.12 4.26 3.86 3.63 3.48 3.37 3.29 3.23 3.18 3.14 3.10 3.07
10.6 8.02 6.99 6.42 6.06 5.80 5.61 5.47 5.35 5.26 5.18 5.11
1.49 1.60 1.60 1.59 1.59 1.58 1.57 1.56 1.56 1.55 1.55 1.54
10 3.28 2.92 2.73 2.61 2.52 2.46 2.41 2.38 2.35 2.32 2.30 2.28
4.96 4.10 3.71 3.48 3.33 3.22 3.14 3.07 3.02 2.98 2.94 2.91
10.0 7.56 6.55 5.99 5.64 5.39 5.20 5.06 4.94 4.85 4.77 4.71
1.47 1.58 1.58 1.57 1.56 1.55 1.54 1.53 1.53 1.52 1.52 1;51
11 3.23 2 86 2.66 2.54 2.45 2.39 2.34 2.30 2.27 2.25 2.23 2.21
4.84 3.98 3.59 3.36 3.20 3.09 3.01 2.95 2.90 2.85 2.82 2.79
9.65 7.21 6.22 5.67 5.32 5.07 4.89 4.74 4.63 4.54 4.46 4.40
1.46 1.56 1.56 1.55 1.54 1.53 1.52 1.51 1.51 1.50 1.50 1.49
12 3.18 2.81 2.61 2.48 2.39 2.33 2.28 2.24 2.21 2.19 2.17 2.15
4.75 3.89 3.49 3.26 3.11 3.00 2.91 2.85 2.80 2.75 2.72 2.69
9.33 6.93 5.95 5.41 5.06 4.82 4.64 4.50 4.39 4.30 4.22 4.16
384
Table D {Continued)
df for numerator
1 -a df for
15 20 24 30 40 50 60 100 120 200 500 00 denom.
9.49 9.58 9.63 9.67 9.71 9.74 9.76 9.78 9.80 9.82 9.84 9.85 .75
61.2 61.7 62.0 62.3 62.5 62.7 62.8 63.0 63.1 63.2 63.3 63.3 .90 1
246 248 249 250 251 252 252 253 253 254 254 254 .95
3.41 3.43 3.43 3.44 3.45 3.45 3.46 3.47 3.47 3.48 3.48 3.48 .75
9.42 9.44 9.45 9.46 9.47 9.47 9.47 9.48 9.48 9.49 9.49 9.49 .90 2
19.4 19.4 19.5 19.5 19.5 19.5 19.5 19.5 19.5 19.5 19.5 19.5 .95
99.4 99.4 99.5 99.5 99.5 99.5 99.5 99.5 99.5 99.5 99.5 99.5 .99
2.46 2.46 2.46 2.47 2.47 2.47 2.47 2.47 2.47 2.47 2.47 2.47 .75
5.20 5.18 5.18 5.17 5.16 5.15 5.15 5.14 5.14 5.14 5.14 5.13 .90 3
8.70 8.66 8.64 8.62 8.59 8.58 8.57 8.55 8.55 8.54 8.53 8.53 .95
26.9 26.7 26.6 26.5 26.4 26.4 26.3 26.2 26.2 26.2 26.1 26.1 .99
2.08 2.08 2.08 2.08 2.08 2.08 2.08 2.08 2.08 2.08 2.08 2.08 .75
3.87 3.84 3.83 3.82 3.80 3.80 3.79 3.78 3.78 3.77 3.76 3.76 .90
5.86 5.80 5.77 5.75 5.72 5.70 5.69 5.66 5.66 5.65. 5.64 5.63 .95 4
14.2 14.0 13.9 13.8 13.7 13.7 13.7 13.6 13.6 13.5 13.5 13.5 .99
1.89 1.88 1.88 1.88 1.88 1.88 1.87 1.87 1.87 1.87 1.87 1.87 .75
3.24 3.21 3.19 3.17 3.16 3.15 3.14 3.13 3.12 3.12 3.11 3.10 .90 5
4.62 4.56 4.53 4.50 4.46 4.44 4.43 4.41 4.40 4.39 4.37 4.36 .95
9.72 9.55 9.47 9.38 9.29 9.24 9.20 9.13 9.11 9.08 9.04 9.02 .99
1.76 1.76 1.75 1.75 1.75 1.75 1.74 1.74 1.74 1.74 1.74 1.74 .75
2.87 2.84 2.82 2.80 2.78 2.77 2.76 2.75 2.74 2.73 2.73 2.72 .90 6
3.94 3.87 3.84 3.81 3.77 3.75 3.74 3.71 3.70 3.69 3.68 3.67 .95
7.56 7.40 7.31 7.23 7.14 7.09 7.06 6.99 6.97 6.93 6,90 6.88 .99
1.68 1.67 1.67 1.66 1.66 1.66 1.65 1.65 1.65 1.65 1.65 1.65 .75
2.63 2.59 2.58 2.56 2.54 2.52 2.51 2.50 2.49 2.48 2.48 2.47 .90 7
3.51 3.44 3.41 3.38 3.34 3.32 3.30 3.27 3.27 3.25 3.24 3.23 .95
6.31 6.16 6.07 5.99 5.91 5.86 5.82 5.75 5.74 5.70 5.67 5.65 .99
1.62 1.61 1.60 1.60 1.59 1.59 1.59 1.58 1.58 1.58 1.58 1.58 .75
2.46 2.42 2.40 2.38 2.36 2.35 2.34 2.32 2.32 2.31 2.30 2.29 .90 8
3.22 3.15 3.12 3.08 3.04 3.02 3.01 2.97 2.97 2.95 2.94 2.93 .95
5.52 5.36 5.28 5.20 5.12 5.07 5.03 4.96 4.95 4.91 4.88 4.86 .99
1.57 1.56 1.56 1.55 1.55 1.54 1.54 1.53 1.53 1.53 1.53 1.53 .75
2.34 2.30 2.28 2.25 2.23 2.22 2.21 2.19 2.18 2.17 2.17 2.16 .90 9
3.01 2.94 2.90 2.86 2.83 2.80 2.79 2.76 2.75 2.73 2.72 2.71 .95
4.96 4.81 4.73 4.65 4.57 4.52 4.48 4.42 4.40 4.36 4.33 4.31 .99
1.53 1.52 1.52 1.51 1.51 1.50 1.50 1.49 1.49 1.49 1.48 1.48 .75
2.24 2.20 2.18 2.16 2.13 2.12 2.11 2.09 2.08 2.07 2.06 2.06 .90 10
2.85 2.77 2.74 2.70 2.66 2.64 2.62 2.59 2.58 2.56 2.55 2.54 .95
4.56 4.41 4.33 4.25 4.17 4.12 4.08 4.01 4.00 3.96 3.93 3.91 .99
1.50 1.49 1.49 1.48 1.47 1.47 1.47 1.46 1.46 1.46 1.45 1.45 .75
2.17 2.12 2.10 2.08 2.05 2.04 2.03 2.00 2.00 1.99 1.98 1.97 .90 11
2.72 2.65 2.61 2.57 2.53 2.51 2.49 2.46 2.45 2.43 2.42 2.40 .95
4.25 4.10 4.02 3.94 3.86 3.81 3.78 3.71 3.69 3.66 3.62 3.60 .99
1.48 1.47 1.46 1.45 1.45 1.44 1.44 1.43 1.43 1.43 1.42 1.42 .75
2.10 2.06 2.04 2.01 1.99 1.97 1.96 1.94 1.93 1.92 1.91 1.90 .90 12
2.62 2.54 2.51 2.47 2.43 2.40 2.38 2.35 2.34 2.32 2.31 2.30 .95
4.01 3.86 3.78 3.70 3.62 3.57 3.54 3.47 3.45 3.41 3.38 3.36 .99
385
df for numerator
1 2 3 4 5 6 7 8 9 10 11 12
1.45 1.54 1.54 1.53 1.52 1.51 1.50 1.49 1.49 1.48 1.47 1.47
13 3.14 2.76 2.56 2.43 2.35 2.28 2.23 2.20 2.16 2.14 2.12 2.10
4.67 3.81 3.41 3.18 3.03 2.92 2.83 2.77 2.71 2.67 2.63 2.60
9.07 6.70 5.74 5.21 4.86 4.62 4.44 4.30 4.19 4.10 4.02 3.96
1.44 1.53 1.53 1.52 1.51 1.50 1.48 1.48 1.47 1.46 1.46 1.45
14 3.10 2.73 2.52 2.39 2.31 2.24 2.19 2.15 2.12 2.10 2.08 2.05
4.60 3.74 3.34 3.11 2.96 2.85 2.76 2.70 2.65 2.60 2.57 2.53
8.86 6.51 5.56 5.04 4.69 4.46 4.28 4.14 4.03 3.94 3.86 3.80
1.43 1.52 1.52 1.51 1.49 1.48 1.47 1.46 1.46 1.45 1.44 1.44
15 3.07 2.70 2.49 2.36 2.27 2.21 2.16 2.12 2.09 2.06 2.04 2.02
4.54 3.68 3.29 3.06 2.90 2.79 2.71 2.64 2.59 2.54 2.51 2.48
8.68 6.36 5.42 4.89 4.56 4.32 4.14 4.00 3.89 3.80 3.73 3.67
1.42 1.51 1.51 1.50 1.48 1.48 1.47 1.46 1.45 1.45 1.44 1.44
16 3.05 2.67 2.46 2.33 2.24 2.18 2.13 2.09 2.06 2.03 2.01 1.99
4.49 3.63 3.24 3.01 2.85 2.74 2.66 2.59 2.54 2.49 2.46 2.42
8.53 6.23 5.29 4.77 4.44 4.20 4.03 3.89 3.78 3.69 3.62 3.55
1.42 1.51 1.50 1.49 1.47 1.46 1.45 1.44 1.43 1.43 1.42 1.41
17 3.03 2.64 2.44 2.31 2.22 2.15 2.10 2.06 2.03 2.00 1.98 1.96
4.45 3.59 3.20 2.96 2.81 2.70 2.61 2.55 2.49 2.45 2.41 2.38
8.40 6.11 5.18 4.67 4.34 4.10 3.93 3.79 3.68 3.59 3.52 3.46
1.41 1.50 1.49 1.48 1.46 1.45 1.44 1.43 1.42 1.42 1.41 1.40
18 3.01 2.62 2.42 2.29 2.20 2.13 2.08 2.04 2.00 1.98 1.96 1.93
4.41 3.55 3.16 2.93 2.77 2.66 2.58 2.51 2.46 2.41 2.37 2.34
8.29 6.01 5.09 4.58 4.25 4.01 3.84 3.71 3.60 3.51 3.43 3.37
1.41 1.49 1.49 1.47 1.46 1.44 1.43 1.42 1.41 1.41 1.40 1.40
19 2.99 2.61 2.40 2.27 2.18 2.11 2.06 2.02 1.98 1.96 1.94 1.91
4.38 3.52 3.13 2.90 2.74 2.63 2.54 2.48 2.42 2.38 2.34 2.31
8.18 5.93 5.01 4.50 4.17 3.94 3.77 3.63 3.52 3.43 3.36 3.30
1.40 1.49 1.48 1.46 1.45 1.44 1.42 1.42 1.41 1.40 1.39 1.39
20 2.97 2.59 2.38 2.25 2.16 2.09 2.04 2.00 1.96 1.94 1.92 1.89
4.35 3.49 3.10 2.87 2.71 2.60 2.51 2.45 2.39 2.35 2.31 2.28
8.10 5.85 4.94 4.43 4.10 3.87 3.70 3.56 3.46 3.37 3.29 3.23
1.40 1.48 1.47 1.45 1.44 1.42 1.41 1.40 1.39 1.39 1.38 1.37
22 2.95 2.56 2.35 2.22 2.13 2.06 2.01 1.97 1.93 1.90 1.88 1.86
4.30 3.44 3.05 2.82 2.66 2.55 2.46 2.40 2.34 2.30 2.26 2.23
7.95 5.72 4.82 4.31 3.99 3.76 3.59 3.45 3.35 3.26 3.18 3.12
1.39 1.47 1.46 1.44 1.43 1.41 1.40 1.39 1.38 1.38 1.37 1.36
24 2.93 2.54 2.33 2.19 2.10 2.04 1.98 1.94 1.91 1.88 1.85 1.83
4.26 3.40 3.01 2.78 2.62 2.51 2.42 2.36 2.30 2.25 2.21 2.18
7.82 5.61 4.72 4.22 3.90 3.67 3.50 3.36 3.26 3.17 3.09 3.03
1.38 1.46 1.45 1.44 1.42 1.41 1.40 1.39 1.37 1.37 1.36 1.35
26 2.91 2.52 2.31 2.17 2.08 2.01 1.96 1.92 1.88 1.86 1.84 1.81
4.23 3.37 2.98 2.74 2.59 2.47 2.39 2.32 2.27 2.22 2.18 2.15
7.72 5.53 4.64 4.14 3.82 3.59 3.42 3.29 3.18 3.09 3.02 2.96
1.38 1.46 1.45 1.43 1.41 1.40 1.39 1.38 1.37 1.36 1.35 1.34
28 2.89 2.50 2.29 2.16 2.06 2.00 1.94 1.90 1.87 1.84 1.81 1.79
4.20 3.34 2.95 2.71 2.56 2.45 2.36 2.29 2.24 2.19 2.15 2.12
7.64 5.45 4.57 4.07 3.75 3.53 3.36 3.23 3.12 3.03 2.96 2.90
Statistical tables 387
Table D {Continued)
df for numerator
1 —a df for
15 20 24 30 40 50 60 100 120 200 500 00 denom.
1.46 1.45 1.44 1.43 1.42 1.42 1.42 1.41 1.41 1.40 1.40 1.40 .75
2.05 2.01 1.98 1.96 1.93 1.92 1.90 1.88 1.88 1.86 1.85 1.85 .90 13
2.53 2.46 2.42 2.38 2.34 2.31 2.30 2.26 2.25 2.23 2.22 2.21 .95
3.82 3.66 3.59 3.51 3.43 3.38 3.34 3.27 3.25 3.22 3.19 3.17 .99
1.44 1.43 1.42 1.41 1.41 1.40 1.40 1.39 1.39 1.39 1.38 1.38 .75
2.01 1.96 1.94 1.91 1.89 1.87 1.86 1.83 1.83 1.82 1.80 1.80 .90
2.46 2.39 2.35 2.31 2.27 2.24 2.22 2.19 2.18 2.16 2.14 2.13 .95 14
3.66 3.51 3.43 3.35 3.27 3.22 3.18 3.11 3.09 3.06 3.03 3.00 .99
1.43 !.41 1.41 1.40 1.39 1.39 1.38 1.38 1.37 1.37 1.36 1.36 .75
1.97 1.92 1.90 1.87 1.85 1.83 1.82 1.79 1.79 1.77 1.76 1.76 .90 15
2.40 2.33 2.29 2.25 2.20 2.18 2.16 2.12 2.11 2.10 2.08 2.07 .95
3.52 3.37 3.29 3.21 3.13 3.08 3.05 2.98 2.96 2.92 2.89 2.87 .99
1.41 1.40 1.39 1.38 1.37 1.37 1.36 1.36 1.35 1.35 1.34 1.34 .75
1.94 1.89 1.87 1.84 1.81 1.79 1.78 1.76 1.75 1.74 1.73 1.72 .90 16
2.35 2.28 2.24 2.19 2.15 2.12 2.11 2.07 2.06 2.04 2.02 2.01 .95
3.41 3.26 3.18 3.10 3.02 2.97 2.93 2.86 2.84 2.81 2.78 2.75 .99
1.40 1.39 1.38 1.37 1.36 1.35 1.35 1.34 1.34 1.34 1.33 1.33 .75
1.91 1.86 1.84 1.81 1.78 1.76 1.75 1.73 1.72 1.71 1.69 1.69 .90 17
2.31 2.23 2.19 2.15 2.10 2.08 2.06 2.02 2.01 1.99 1.97 1.96 .95
3.31 3.16 3.08 3.00 2.92 2.87 2.83 2.76 2.75 2.71 2.68 2.65 .99
1.39 1.38 1.37 1.36 1.35 1.34 1.34 1.33 1.33 1.32 1.32 1.32 .75
1.89 1.84 1.81 1.78 1.75 1.74 1.72 1.70 1.69 1.68 1.67 1.66 .90 18
2.27 2.19 2.15 2.11 2.06 2.04 2.02 1.98 1.97 1.95 1.93 1.92 .95
3.23 3.08 3.00 2.92 2.84 2.78 2.75 2.68 2.66 2.62 2.59 2.57 .99
1.38 1.37 1.36 1.35 1.34 1.33 1.33 1.32 1.32 1.31 1.31 1.30 .75
1.86 1.81 1.79 1.76 1.73 1.71 1.70 1.67 1.67 1.65 1.64 1.63 .90 19
2.23 2.16 2.11 2.07 2.03 2.00 1.98 1.94 1.93 1.91 1.89 1.88 .95
3.15 3.00 2.92 2.84 2.76 2.71 2.67 2.60 2.58 2.55 2.51 2.49 .99
1.37 1.36 1.35 1.34 1.33 1.33 1.32 1.31 1.31 1.30 1.30 1.29 .75
1.84 1.79 1.77 1.74 1.71 1.69 1.68 1.65 1.64 1.63 1.62 1.61 .90 20
2.20 2.12 2.08 2.04 1.99 1.97 1.95 1.91 1.90 1.88 1.86 1.84 .95
3.09 2.94 2.86 2.78 2.69 2.64 2.61 2.54 2.52 2.48 2.44 2.42 .99
/
1.36 1.34 1.33 1.32 1.31 1.31 1.30 1.30 1.30 1.29 1.29 1.28 .75
1.81 1.76 1.73 1.70 1.67 1.65 1.64 1.61 1.60 1.59 1.58 1.57 .90 22
2.15 2.07 2.03 1.98 1.94 1.91 1.89 1.85 1.84 1.82 1.80 1.78 .95
2.98 2.83 175 2.67 2.58 2.53 2.50 2.42 2.40 2.36 2.33 2.31 .99
1.35 1.33 1.32 1.31 1.30 1.29 1.29 1.28 1.28 1.27 1.27 1.26 .75
1.78 1.73 t70 1.67 1.64 1.62 1.61 1.58 1.57 1.56 1.54 1.53 .90 24
2.11 2.03 1.98 1.94 1.89 1.86 1.84 1.80 1.79 1.77 1.75 1.73 .95
2.89 2.74 2.66 2.58 2.49 2.44 2.40 2.33 2.31 2.27 2.24 2.21 .99
1.34 1.32 1.31 1.30 1.29 1.28 1.28 1.26 1.26 1.26 1.25 1.25 .75
1.76 1.71 1.68 1.65 1.61 1.59 1.58 1.55 1.54 1.53 1.51 1.50 .90 26
2.07 1.99 l95 1.90 1.85 1.82 1.80 1.76 1.75 1.73 1.71 1.69 .95
2.81 2.66 2.58 2.50 2.42 2.36 2.33 2.25 2.23 2.19 2.16 2.13 .99
1
1.33 1.31 1.30 1.29 1.28 1.27 1.27 1.26 1.25 1.25 1.24 1.24 .75
1.74 1.69 1.66 1.63 1.59 1.57 1.56 1.53 1.52 1.50 1.49 1.48 .90 28
2.04 1.96 1.91 1.87 1.82 1.79 1.77 1.73 1.71 1.69 1.67 1.65 .95
2.75 2.60 2.52 2.44 2.35 2.30 2.26 2.19 2.17 2.13 2.09 2.06 .99
388 Statistical tables
Table D {Continued)
df for numerator
df for 1 —a
denom. 1 2 3 4 5 6 7 8 9 10 11 12
.75 1.38 1.45 1.44 1.42 1.41 1.39 1.38 1.37 1.36 1.35 1.35 1.34
30 .90 2.88 2.49 2.28 2.14 2.05 1.98 1.93 1.88 1.85 1.82 1.79 1.77
.95 4.17 3.32 2.92 2.69 2.53 2.42 2.33 2.27 2.21 2.16 2.13 2.09
.99 7.56 5.39 4.51 4.02 3.70 3.47 3.30 3.17 3.07 2.98 2.91 2.84
.75 1.36 1.44 1.42 1.40 1.39 1.37 1.36 1.35 1.34 1.33 1.32 1.31
40 .90 2.84 2.44 2.23 2.09 2.00 1.93 1.87 1.83 1.79 1.76 1.73 1.71
.95 4.08 3.23 2.84 2.61 2.45 2.34 2.25 2.18 2.12 2.08 2.04 2.00
.99 7.31 5.18 4.31 3.83 3.51 3.29 3.12 2.99 2.89 2.80 2.73 2.66
.75 1.35 1.42 1.41 1.38 1.37 1.35 1.33 1.32 1.31 1.30 1.29 1.29
60 .90 2.79 2.39 2.18 2.04 1.95 1.87 1.82 1.77 1.74 1.71 1.68 1.66
.95 4.00 3.15 2.76 2.53 2.37 2.25 2.17 2.10 2.04 1.99 1.95 1.92
.99 7.08 4.98 4.13 3.65 3.34 3.12 2.95 2.82 2.72 2.63 2.56 2.50
.75 1.34 1.40 1.39 1.37 1.35 1.33 1.31 1.30 1.29 1.28 1.27 1.26
120 .90 2.75 2.35 2.13 1.99 1.90 1.82 1.77 1.72 1.68 1.65 1.62 1.60
.95 3.92 3.07 2.68 2.45 2.29 2.17 2.09 2.02 1.96 1.91 1.87 1.83
.99 6.85 4.79 3.95 3.48 3.17 2.96 2.79 2.66 2.56 2.47 2.40 2.34
.75 1.33 1.39 1.38 1.36 1.34 1.32 1.31 1.29 1.28 1.27 1.26 1.25
200 .90 2.73 2.33 2.11 1.97 1.88 1.80 1.75 1.70 1.66 1.63 1.60 1.57
.95 3.89 3.04 2.65 2.42 2.26 2.14 2.06 1.98 1.93 1.88 1.84 1.80
.99 6.76 4.71 3.88 3.41 3.11 2.89 2.73 2.60 2.50 2.41 2.34 2.27
.75 1.32 1.39 1.37 1.35 1.33 1.31 1.29 1.28 1.27 1.25 1.24 1.24
00 .90 2.71 2.30 2.08 1.94 1.85 1.77 1.72 1.67 1.63 1.60 1.57 1.55
.95 3.84 3.00 2.60 2.37 2.21 2.10 2.01 1.94 1.88 1.83 1.79 1.75
.99 6.63 4.61 3.78 3.32 3.02 2.80 2.64 2.51 2.41 2.32 2.25 2.18
Statistical tables 389
Table D* {Continued)
df for numerator
df for
15 20 24 30 40 50 60 100 120 200 500 00 1 — a denom.
1.32 1.30 1.29 1.28 1.27 1.26 1.26 1.25 1.24 1.24 1.23 1.23 .75
1.72 1.67 1.64 1.61 1.57 1.55 1.54 1.51 1.50 1.48 1.47 1.46 .90 30
2.01 1.93 1.89 1.84 1.79 1.76 1.74 1.70 1.68 1.66 1.64 1.62 .95
2.70 2.55 2.47 2.39 2.30 2.25 2.21 2.13 2.11 2.07 2.03 2.01 .99
1.30 1.28 1.26 1.25 1.24 1.23 1.22 1.21 1.21 1.20 1.19 1.19 .75
1.66 1.61 1.57 1.54 1.51 1.48 1.47 1.43 1.42 1.41 1.39 1.38 .90 40
1.92 1.84 1.79 1.74 1.69 1.66 1.64 1.59 1.58 1.55 1.53 1.51 .95
2.52 2.37 2.29 2.20 2.11 2.06 2.02 1.94 1.92 1.87 1.83 1.80 .99
1.27 1.25 1.24 1.22 1.21 1.20 1.19 1.17 1.17 1.16 1.15 1.15 .75
1.60 1.54 1.51 1.48 1.44 1.41 1.40 1.36 1.35 1.33 1.31 1.29 .90 60
1.84 1.75 1.70 1.65 1.59 1.56 1.53 1.48 1.47 1.44 1.41 1.39 .95
2.35 2.20 2.12 2.03 1.94 1.88 1.84 1.75 1.73 1.68 1.63 1.60 .99
1.24 1.22 1.21 1.19 1.18 1.17 1.16 1.14 1.13 1.12 1.11 1.10 .75
1.55 1.48 1.45 1.41 1.37 1.34 1.32 1.27 1.26 1.24 1.21 1.19 .90 120
1.75 1.66 1.61 1.55 1.50 1.46 1.43 1.37 1.35 1.32 1.28 1.25 .95
2.19 2.03 1.95 1.86 1.76 1.70 1.66 1.56 1.53 1.48 1.42 1.38 .99
1.23 1.21 1.20 1.18 1.16 1.14 1.12 1.11 1.10 1.09 1.08 1.06 .75
1.52 1.46 1.42 1.38 1.34 1.31 1.28 1.24 1.22 1.20 1.17 1.14 .90 200
1.72 1.62 1.57 1.52 1.46 1.41 1.39 1.32 1.29 1.26 1.22 1.19 .95
2.13 1.97 1.89 1.79 1.69 1.63 1.58 1.48 1.44 1.39 1.33 1.28 .99
1.22 1.19 1.18 1.16 1.14 1.13 1.12 1.09 1.08 1.07 1.04 1.00 .75
1.49 1.42 1.38 1.34 1.30 1.26 1.24 1.18 1.17 1.13 1.08 1.00 .90 00
1.67 1.57 1.52 1.46 1.39 1.35 1.32 1.24 1.22 1.17 1.11 1.00 .95
2.04 1.88 1.79 1.70 1.59 1.52 1.47 1.36 1.32 1.25 1.15 1.00 .99
* Abridged from E. S. Pearson and H. Hartley (Eds.), Biometrika Tables for Statisticians,
vol. 1 (2nd Ed.). New York; Cambridge University Press (1958), Table 18. Reproduced by
permission of E. S. Pearson and the trustees of Biometrika.
r- Tj- m ON NO r~- to to ,—1 Y—1 to NO O ON lo T—1 ON OO NO ■tt to
o NO r- q (N to y—i q NO ■tj- to fN t-H O ON ON OO q r- q to q q q
NO q td id id
(N ON od K K NO NO N6 NO NO NO NO td td td td td id td id id to
ON
T™^ ON
fj
.j r- (N ■ri- OO O OO l-H to NO 00 ON O NO to to <N o ON r- ON
ON
00 m l-H q to q OO to t;}- (N i-H O ON OO r' q
q q q Tj- to (N q ON to
6
od NO' q ON od t--' r-' NO NO’ NO' NO NO’ NO td td td td id id id td td td td id q .
IT)
"o ci
V) >
NO OO m to m ro .-Htt o ON o fN to ON Tj- ON lO O 00 to q to
OC q ON O q q q to to q o o ON OO q q q NO q q to q q q q _E
od NO' d ON K r-' NO NO NO NO NO NO id id td id id id td td id td td id Tj- 4—»
r;. •r^ TH c/3
<U
!U C*H
OO to (N (N ■ttto ON ON to O o (N to ON NO to to to Tj- q T3 o
q
i
Tj- m ro ro to to q NO <o q q q q q q q to ■q ro cq <N q q oq q t/5
(U
VN d od fd NO NO NO NO td td td id id id td td id td id td td Tj- Tj-’ q’ (U
\
a> GO
'CS x: (U
] OO NO r-- ON (N ON ON to to NO ON ■tj- ON to tN 00 00 OO OO 00 OO OO TD
m (N q q m ON to q q q OO q q q ■q to to to q q q q oq q NO
V o o
rd td o od K NO NO NO NO td td td id id id td id td id id id id q on
X
+-*
>o
_C _C/5
? (N 0\
to to to ON (N OO
q ON q q q tt q
OO
q
to
OO q q
to
to
NO o to
q
to O O O
to CO <N q <N q
o
q oq q q
(N 'o
s;
<N
, PH
■'^" ON od r^' NO' NO' NO to td td td td td td id td td id id td td •rf •'tf q <u ,
* lo qj
* _>
r- t-- q ■rt lO 3
ON (N <o o to ON to tN to NO NO (N (N co g o
NO m q q q (N q OO q q q q to to q q q q q q OO q q to 0) >
d ON od NO' NO NO to’ td td td id id id td td id id lo id q ■^’ q'
C_)
_c
m (U
OH (/5
Vo
NO m m ON
ON to (N o ■tt ON O (N to O to TO <N to Tj- to NO i-,
tu
q 00 ON q q q q to q q q q q q q q to q 0)
10
q OO GO
"H*
ON rd ON r-‘ NO NO
a
NO to to' td td td td td td id td id id td TJ-' q q’ a
c
*
VH
D
0)
d) OO o <N ON o NO to ON to OO CO ON NO <N O <N to to ON <u .O </3
bfi q to q q m q q q q to q q q q q q q oq q q q q to x; • •»«» o
■'tf
ON »
C r-‘ rd ON r-‘ NO NO' to’ to’ to’ td td td id td td td ''ti- q q q' o:) X
03
o CHH
o
c/5
cn to to <N rs| O o to O O CN to ON ■tt" O ND to ON OO o <N NO ON (U
c/5 c/5
<u OO
q q m to OO q q to q q q q q q 00 oq q q q q q q cq q 3
<U
(U
c
o3
_N iri rd od r~' NO NO to’ to id td td to td ■t;}- -td T}-' q’ 03 <u
-4->
c
H
C/5
P
VH
s
T3
d) m t^- NO OO ON ON o ■ttt (N to to OO to OO o (N NO ON q r- 0) (U
(U
T3
r- ’'t 'tt o (N q to q q q q OO OO q q q q q q lO q to cq q HH
ax:
O
3 fd <N’ od r-’ NO to' to' to td id id ■tj-’ q ■tj- ■tt ■rj- "tt ''tf ti- q’ o p
GO c*-
u. o c3
O X
o m ro ON m to (N ^-H (N to ON ■'tj" ON NO tN ON to r-. o to o c (L>
>o q o q q q CO q q q OO q q q to to to q q q to q q
NO to
q U _o u.
O) NO to’ td ■';}■ ■rj- ■rt- TO
d »-H od NO to to’ to’ q q "tt’ q' c c/5 3
o3 'H
c/5
ON .-H
■td- NO ON NO NO OO NO _
ooNO ■ttj- ^H ttl- OO (N NO T3 c 1.-
OH
(N 00 to CO q ro q OO q q to' to q to to to q q q q ON q OO
<U (U +->
TD a
-(->
C
r-' d r-‘ NO to to to ■'ti- ■tj- 'tt ■'tj- ■tti-' rj- rd rd rd c c
d)
m (U 03
d) X
X) 3
o ro NO OO Q OO to tN to NO o to OO to tN O OO NO o Tf ON ON CO UJ 'T3 cr
d)
OO OO OO q q q to q to q q q q q q q q q q q q q q (U CHH
OH
to ■rj- Tt- q •r}- O O
(N
fO
ON NO to' "tt rd rd rd rd rd td rd td 3
»o >T TO VH
o3 o (U
X
d) OO OO o NO to OO (N r-- to o r-- to tN ON OO to OO o NO tN S SH a c
cx q q q q q q q q q q q q q q q q c
26.7
OO to ro OO OO to to to
a r<^ od to* to’ rd rd rd rd rd rd rd rd rd td rd rd rd rd rd rd rd rd
(U 3
0^ C
D
O
rd
ON o <o NO NO O to OO NO to O OO NO to (N ON NO to o ON X
p q to q q q ro q q q q q p q q. ON q ON ON q OO oq OO r- £ t/5
(N 7
U OO NO •rt- rd ro CO rd rd rd rd to rd rd rd rd tN (N IN tN tN (N tN tN (N tN o CN
o ON
mm iJs
*
<N
1—i (N m to NO r- OO ON O <N to q NO OO ON O O O O o o * *
a tN (N to •t;!- NO tN o a.
' '
H \ CL.
390
c
CO VO CO CM VO CO VO Ov VO VO VO VO oo CM T-H T-H T—H CM CO VO
Ov oo q VO VO O q q Ov c- VO CO CM T-H o GO oo OO VO CM q 00 q
20
00 r~' Ov T-H d Ov Ov od od r^’ cd cd cd cd cd cd vd vd vd vd vd vd vd vd vd
o^ m T-H T-H
(N
T-H CO VO ov VO oo VO oo CO o OV O VO VO VO C' oo ov
CM OO VO q T-H oo VO CO CM o O oo oo c- q CO T—H q q VO
K Ov T-H d Ov od od od u cd cd cd cd cd cd vd vd vd vd vd vd vd vd vd
Os T-H T-H
p is the number of quantities (for example, means) whose range is involved, tiz is the degrees of freedom in the error estimate.
(N
VO VO VO CM ov CO CM o C- c- c- oo CM VO VO VO Ov T-H
IT) ov VO CM CO q CO ov VO CO CM o q oo C' C^ VO CM q OO q q
'O vd Ov cd T—i d Ov od Ov U c-’ cd cd cd cd vd vd vd vd vd vd vd vd vd vd vd
oo m T-H T-H T-H T-H
(N
o oo Tt VO CO VO r-- CM o o o CM VO oo oo o CM Tj- VO Ov
q q q o CM q q Ov VO CM T-< o Ov oo C' VO VO CO CM O q q
vd od cd T-H d ov od od U K cd cd cd cd vd vd vd vd vd vd vd vd vd vd vd
00 T-H T-H
fN
VO CM VO CO VO VO Ov VO CO CM CO VO oo CM CO VO ov VO
m VO CM ov T—< q q oo VO CO T-H o Ov OO c- VO VO VO CO oo q
r-- lO oo' cd T-H Ov Ov od od c-’ u cd cd cd vd vd vd vd vd vd vd vd vd vd vd vd
(N m T-H
rsl
oo T-H o CO VO VO o VO VO oo T-H VO VO 00 o CO VO O
q <N CO q q o q; O c~ CM T—1 Ov oo VO VO VO TC CM o q q
oo’ cd HH Ov Ov od od u C-’ cd cd .vd vd vd vd vd vd vd vd vd vd vd vd vd
r-- ro T-H
(N
o Ov T-H oo oo oo VO VO o c- VO VO 00 oo CO c- O Ov
lO oo q c- CM O Ov C' VO VO TC CO CM 1-H q q q Tj- q
O fd rd CM’ d ov od od K C-’ K cd vd vd vd vd vd vd vo’ vd vd vd vd vd vd vd
MD m T-H T-H T-H
CM
oo o VO CO VO VO CO ov VO V) VO oo VO oo CM VO oo CO oo CO
VO q CO VO O VO CO T—1 q c- VO q CO CO CM O q q VO CO CM
m r4 r--’ CM’ d ov od od c-’ u cd vd vd vd vd vd vd vd vd vd vd vd vd vd vd vd
IT) T-H
CM
O c- c- ov T-H ov T-H n- VO c- o oo CM VO o VO O VO
r-; r- CO CM CO 00 CM Ov 00 VO VO CO CM CM o OO c- VO CO q
o
VO 1-H ■sd CM’ d Ov od C-’ u r-‘ vd vd vd vd vd vd vd vd vd vd vd vd vd vd vd vd
T-H
Table E.2 Upper 1 Percent Points of Studentized Range q
(N
r- 00 CM VO c- CO T-H CM VO oo CM C" VO o VO 00
CM q Ov oo T-H VO CO O oo q VO CO CM T-H o o oo oo q q CO q o
ON
r- o vd HH Ov od od C-’ vd vd vd vd vd vd vd vd vd vd vd vd vd vd vd vd
m cn r—( T-H
(N
c- CO c- c- VO VO oo Tl- oo TC oo Ov VO CM Ov
m VO q VO q Ov Tj- q 00 MD VO CO CM T-H o o q q oo q VO CO CM T-H q
00
Ov vd T-H Ov od r--’ U vd vd vd vd vd vd vd vd vd vd vd vd vd vd vd vd
CM <N T-H
CM
CM CM 00 00 CM ov oo Ov CM VO Oo CO oo o CO oo
CM q T—^ CO CO VO q ov q q o q Ov q q q q VO ■q q T-H q q
VO oo’ vd T—T ov od U U vd vd vd vd vd vd vd vd vd vd vd vd vd vd vd vd vd
<N r-H T-H
CM
VO CM VO CO VO c- CO CO VO Ov CO oo CO oo C- VO CO CM T-H o
t-- CO ov q O VO CO q q q q VO CO q q q q q q q
VO cd OV od U U vd vd vd vd vd vd vd vd vd vd vd vd vd vd vd
oo CM
C' o CO o VO c- CM O o CM VO Ov TC oo VO CM O o O o o
m CM T-H q O VO CM q q VO VO q CO q T-H T-H q o O q q q q q •q
CM CM ov u u vd vd vd vd vd vd vd vd vd vd vd vd vd vd ■rr
VO CM
T—(
CM c- CO CM CO CO c- 'Tj- VO oo CO oo TC O C- TC VO oo O CM
O VO T-H ov CO Ov q q T-H q q q q q q q q q VO q q q T-H
rn
V-) OV d oo’ vd vd vd vd vd vd vd vd Tf
tn
VO o VO O oo Ov CM VO T-H c- CO o V) CM VO Ov CM VO o
o q CM VO r- CM q q q ■q q CO q CM q q q q q q q q q q
(N
o 00 vd vd vd Tj- Tf Tj-’ cd cd cd cd cd cd
OV
T-H rN CO VO VO 00 ov o CM CO VO VO c~ oo oo o Tt- O O O O Q
N T-H T-H T-H T-H T-H T-H T-H TT—1 —H T-H CM CM CO VO CM o
391
392 Statistical tables
k Polynomial 1 2 3 4 5 6 1 8 9 10 Sc,? 2
3 Linear -1 0 1 2 1
Quadratic 1 -2 1 6 3
Linear -3 -1 1 3 20 2
4 Quadratic 1 -1 -1 1 4 1
1 0
Cubic -1 3 -3 1 20 3
Linear -2 -1 0 1 2 10 1
5 Quadratic 2 -1 -2 -1 2 14 1
5
Cubic -1 2 0 -2 1 10 6
3 5
Quartic 1 -4 6 -4 1 70 12
Linear -5 -3 -1 1 3 5 70 2
3
6 Quadratic 5 -1 -4 -4 -1 5 84 2
5
Cubic -5 7 4 -4 -7 5 180 3
7
Quartic 1 -3 2 2 -3 1 28 12
Linear -3 -2 -1 0 1 2 3 28 1
7 Quadratic 5 0 -3 -4 -3 0 5 84 1
Cubic -1 1 1 0 -1 -1 1 6 1
6
Quartic 3 -7 1 6 1 -7 3 154 12
Linear -7 -5 -3 -1 1 3 5 7 168 2
Quadratic 7 1 -3 -5 -5 -3 1 7 168 1
8 Cubic -7 5 7 3 -3 -7 -5 7 264 2
3
Quartic 7 -13 -3 9 9 -3 -13 7 616 7
12
Quintic -7 23 -17 -15 15 17 -23 7 2184 7
1 0
Linear -4 -3 -2 -1 0 1 2 3 4 60 1
Quadratic 28 7 -8 -17 -20 -17 -8 7 28 2772 3
9 Cubic -14 7 13 9 0 -9 -13 -7 14 990 5
6
Quartic 14 -21 -11 9 18 9 -11 -21 14 2002 7
12
Quintic -4 11 -4 -9 0 9 4 -11 4 468 3
20
Linear -9 -7 -5 -3 -1 1 3 5 7 9 330 2
Quadratic 6 2 -1 -3 -4 -4 -3 -1 2 6 132 1
2
10 Cubic -42 14 35 31 12 -12 -31 -35 -14 42 8580 5
3
Quartic 18 -22 -17 3 18 18 3 -17 -22 18 2860 5
12
Quintic -6 14 -1 -11 -6 6 11 1 -14 6 780 1
1 0
Answers to Odd-Numbered Problems
CHAPTER 2
2.1 Y = 18,472.9, = 41.7.
2.3 For a two-sided alternative, some points are
fi = 18,473; 18,472; 18,471; 18,470; 18,469; 18,468; 18,467
= = 0.08, 0.39, 0.80, 0.95, 0.80, 0.39, 0.08.
2.5 n = 1.
2.7 Do not reject hypothesis, since = 17.05 with 11 df.
2.9 Do not reject hypothesis, since F = 1.8.
2.11 Reject hypothesis at 1 percent level, since \t\ = 13.4.
2.13 1. Reject hypothesis, since z = 4.0.
2. Do not reject hypothesis, since z = 1.71.
3. Do not reject hypothesis z, since z < 1.
2.15 For = 13 15 17 19
Power (1 - 13) = 0.04 0.27 0.70 0.95.
2.17 Reject hypothesis at a = 0.05, since t = 8.32.
2.19 Reject hypothesis of equal variances at a = 0.05, since F = 17.4. Do not reject
hypothesis of equal means at a = 0.05, since t' is <1.
2.21 Reject hypothesis at a = 0.05, since t = 2.68.
2.23 Do not reject hypothesis at a = 0.05, since t = 1.62.
2.25 Reject hypothesis at a = 0.05, since z = — 1.96.
CHAPTER 3
3.1
Source df SS MS
Totals 24 329.84
393
394 Answers to odd-numbered problems
SS
= TA - Tc = 31 60.06
C2 = T^- 2T, + TC^ -79 130.02
190.08
3.5
B A C
Yj = 25.4 22.4 18.5
SS
Set 1 : Cl = 2Ti — 275 =
-28 49.0
C2 = 4T2 — 674 =
-108 48.6
C3 = llTi + IIT2 - 1473 + 1174 + 1175 =
71 1.3
C4 = lOTi 4T2
— —
4T4 + 1075 8 0.1
99.0
SS
Set 2 : Cl = 6T1 2T2
— =
-18 3.4
C2 = llTi + 1172 - 873 =
-237 33.6
C3 = 4Ti + 472 + 473 — 1974 = -252 36.3
C4 - 2Ti + 272 + 273 + 2T4 — 2375 = -172 25.7
99.0
3.9 75.6 percent idue tc1 levels of A, 24.4 ]percent due to error.
3.11 Two might be C 1 —^ C.i -- T 2 and C2 = 27.1 - T.2 - C3 and C
which is less than = 60.65, so is not significant at 5 percent, and C2 = 7,
which is less than = 105.07, so is not significant at 5 percent.
3.13
Source df SS MS 7
3.15
Source df SS MS F
3.19 One case based on n — 1 df and the other on Newman-Keuls. The latter is
better if variances are homogeneous.
3.21 One set of orthogonal contrasts might be
Cl = 4T, - T, - 7,3 - - T,
C2 = Ti - T3
C3 = T4 - T,
C4 = Ti + T3 - - Ts
3.23 For in Problem 3.21, = 24.5.
3.25 Litter to litter differences account for 28.6 percent of the total variance.
CHAPTER 4
4.1
Source df SS MS
Totals 11 2.28
Source df SS MS
4.9 Proof
396 Answers to odd-numbered problems
4.11 Error df = 0.
4.13 Three groups (treatments) and six lessons (blocks) ANOVA
Source df
Groups 2
Lessons 5
Error 10
Source df
Scales 4
Times 4
Days 4
Error 12
b. ANOVA;
Source df
Scales 4
Days 4
Error 16
4.19 HQ. Fj ^ 0 for all j. Model: Yfj = Fj + + 8^. F = 8.75, which is signifi-
cant at the 1 percent level.
4.21 From —0.38 to 0.88.
4.23 For example,
Cl = T.i - T,
C2 = T.i + 7.2 - 27.3
C3 = F3 - 7.4
4.25
Source df SS MS F
CHAPTER 5
5.1
Source df SS MS F
Phosphor: CAB
Means: 253.33 260.00 273.33
so A or C is better than B.
5.5
Source df SS MS
Totals 35 52.30
5.9 Plot of cell totals versus feed for the two material lines are not parallel (interaction),
and both material and feed effect are obvious.
5.11
Source df SS MS
Totals 71 610;38
398 Answers to odd-numbered problems
5.13
Source df SS MS
Types: 3 2 1
Means: 3.6 5.2 7.3
Hence type 1 gives best yield with any of the five soil types.
J by A means:
1.04 1.14 1.17 1.20 1.24 1.42 1.58 1.59 1.70 1.70
Source df SS MS
Totals 63 6636.67
Answers to odd-numbered problems 399
5.23 a.
Source df
Waxes 3
Times 2
W X T interaction 6
Error 12
b. Changes in gloss index due to the four waxes are different for the three polishing
times.
CHAPTER 6
Source df SS MS
Factor A 1 0.08 0.08
Factor B 1 70.08 70.08**
A X B interaction 1 24.08 24.08
Error 8 36.68 4.58
Totals 11 130.92
Source df SS MS
A 1 2704.00 2704.00
B 1 26,732.25 26,732.25
AB 1 7744.00 7744.00
C 1 24,025.00 24,025.00
AC 1 16,256.25 16,256.25
BC 1 64,516.00 64,516.00
ABC 1 420.25 420.25
Error 8 246,284.00 30,785.50
Totals 15 388,681.75
6.5 No plots, since there are no significant effects at the 5 percent level.
400 Answers to odd-numbered problems
6.7
Source df SS MS
A 1 13,736.5
B 1 6188.3
AB 1 22,102.5 Same
C 1 22.7 as
AC 1 22,525.0 SS
BC 1 12,051.3
ABC 1 20,757.0
D 1 81,103.8
AD 1 145,665.0*
BD 1 9214.3
ABD 1 126,630.3*
CD 1 148.8
ACD 1 6757.0
BCD 1 294.0
ABCD 1 19,453.8
Error 16 431,599.4 26,975.0
Totals 31 918,249.7
6.9 a. (1), a, b, ab, c, ac, be, abc, d, ad, bd, abd, cd, acd, bed, abed.
b. Source df
Total 15
Source df
Total 31
Answers to odd-numbered problems 401
6.17
Source df SS and MS
6.19
Source df SS and MS
A 1 45.60*
B 1 1.36
AB 1 53.56*
C 1 0.01
AC 1 0.00
BC 1 0.01
ABC 1 0.04
6.21 Plot the following: (1) = 0, a = —3,h— —2,ab = 5,c— —l,ac = 0,bc= 1,
and abc = — 2.
CHAPTER 7
7.13
Source df SS
7.15
Source df SS MS
Linear only on gate setting giving: y[, = 2.517 + 1.15M Oi X = 0.217 + 0.575A,.
7.19
Source df Source df
R 1 G X W 4
G 2 Gz. X 1
GL 1 GL X WQ 1
GQ 1 GQ X W^ 1
R X G 2 GQ X WQ 1
R X Gi^ 1 R X G X W 4
R X GQ 1 R X X ILL 1
W 2 R X GL X WQ 1
1 R X GQ X Wj^ 1
% 1 R X GQ X WQ 1
R X W 2 Error 18
R X 1
R X WQ 1 Total 35
7.21 Graph.
Answers to odd-numbered problems 403
12?>
Source df Source df
Radius (R) 3 RL X FL 1
RL 1 RL X FQ 1
RQ 1 RL X Fc 1
Rc 1 RQ X FL 1
Force (F) 3 RQ X Fa 1
FL 1 RQ X Fc 1
FQ 1 Rc X FL 1
Fc 1 Rc X FQ 1
R X F interaction 9 Rc X Fc 1
Error 32
Total 47
7.25 Plot.
CHAPTER 8
CHAPTER 9
Source df SS MS
Totals 17 1720.44
9.3 AB = 320.111
358.22
AB^ = 38.11J
9.5
Source df SS MS
V 2 2.79 1.40*
T 2 9.01 4.50**
V X T 4 1.99 0.50
Error 9 2.15
9.11 After coding data by subtracting 2.6 and multiplying by 10, we have
Source df SS MS
All except B x C are significant at the 5 percent level, but the main effects pre-
dominate.
9.13
Source df SS
AB 2 66.931
185.19
AB^ 2 118.26)
AC 2 58.931
189.75
AC^ 2 130.82)
BC 2 5.821
27.75
BC^ 2 21.93)
ABC 2 17.603
ABC^ 2 8.26 [
100.05
AB^C 2 53.48
AB^C^ 2 20.71 /
9.15 Plots show strong linear effects of A, B, and C. They show AB and AC inter-
action—but it is slight compared with the main effects. A slight quadratic trend
in A is also noted.
9.17
Source df SS MS
Totals 35 31.530
CHAPTER 10
10.1
Source EMS
(Tg + lOo-J
+ 2a o A + 6^^-
OA,j (T^ +
Wg <— J ^
%ij)
10.3
Source EMS
A + UCGIB + nbcG^
Bj + nca^B + nacai
AB,j + UCGIB
Q + ^^ABC + naaic + nba^c + nab(l)c
AC,, + ^^ABC + nbalc
BCjk + ^^ABC + naalc
+ ^^ABC
Source EMS
Source df EMS
Ti 3 + 2GJB T 10(^E\
Sj 4 (^8 + 8cr| )
TSu 12 <y^
20
Source EMS
0. + +
2OQI 8(TQ
ij (^8 + +
2GQI 8(TJ
0/,, + 2^01
Instruments; 40 percent
Operators: 32 percent
Operator/
instrument interaction; 15 percent
Error; 13 percent
CHAPTER 11
11.1
Source df SS MS EMS
Totals 35 85.60
Source EMS
(^8
Source df SS MS EMS
11.7
Source EMS
Ti + 18cr^ + 36
^Ki) + 18a^
s. + + 36
TSi, + +
^^kj(i) + 9cr^5
Pm + + 24 (/>p
TPim <y^ + ^^mp + l2(/)fp
+ ^^mp
SP,m (^8 +I 3a
^^nxsp + 12<^sp
'3^2
TSPij^^ + ^^msp "b ^^tsp
'3—2
+ -^^msp
^q(ijkm) (^8
11.9 a, b.
Source df EMS
c, 1 + 2ol + 8^2
Tj(i)
6 + 2aj
Vk 1 + ajv + 4ocv
CV,k 1 + Gpy + 4OQV
FKm 6 (^1 + olv
c. 60 percent.
Answers to odd-numbered problems 409
11.11 Newman-Keuls shows experimental posttest group means better than control
and one experimental group better than the other.
11.19
Source df EMS
Di 2 2cr| + llaj)
Gj 1 ^SG + llo-oG + 33^;^G
DG,j 2 ^SG + 11^DG
^k(i) 30 2GI
30 ^SG
11.21
Source df EMS
Ri 1 + 6cri + 36^1
Hjw 10 + 6aH
Pk 2 + 2(JpH + \2opp + 24</>p
RPiK 2 + 2(7pH + I2app
P^kKi) 20 + 2GPH
^m(ijk) 36
11.23
Source df EMS
^q(ijkm) 180
410 Answers to odd-numbered problems
11.25 Recommend 1:1 concentration and any of the fillers—iron filings, iron oxide, or
copper.
CHAPTER 12
12.1
Source df SS MS EMS
Totals 53 1068.59
12.3
Source df SS EMS
R^ik(J) 2 G^ + ^(^RM
^miijk) 16
Total 23
No direct test on types, but results look highly significant. No other signifi-
cant eflfects.
12.5
Source df EMS
Ri 2 lOai No test
4
RM,j 8 ^^RM No test
S, 1 5GIS + 150s\ Poor test
RStk 2 ^^RS No test
MS,, 4 '^RMS + 30MS)
2
8 ^RMS No test
Answers to odd-numbered problems 411
Source df EMS
R, 2 30(7^
yj 4 6(TRy + 18(/)^\
R|/, 8 6GIV ^
5 Sals +
RSik 10 54s ^
VSj, 20 ^RVS + 3</)J/5\
RVS,j, 40 ^RVS
Source df EMS
^1 1 4 + 364
Pi 2 GI + 12(T4 + 24(/)p^ Poor test
RPij 2 + 124P
H, 5 G} + ^4//
RHi, 5 + ^^RH
PHj, 10 + ^^RPH + 4(/)p^'\
RPH,j, 10 + ^^RPH 2F
c 36
^m(ijk) <^c
CHAPTER 13
Source df EMS
Total 35
b. Choose one of the orifice sizes at random, randomize flow rate through this
orifice, then move to a second orifice size, and so on.
Source df
2
Oi 2
RO,j 4
Fk 2
RF,, 4
OFj, 4
ROF iji^ 8
Source df EMS
Whole Ri 3 -f AOal
plot Sj 1 ol + 4cr|5^ + 16cr|^ + 20cr|s + 80 ^5
RS,j 3 + ^^IsA + 20cris
Jk 1 0^ + + 16(TJ^ + IOORJ + ^0(J)j
RSJAiji^^ 12 GE + ^(^RSJA
2
^qiijkm) 80 f^E
Source df MS EMS
Total 59
13.9 In a nested experiment the levels of B, say, are different within each level of A.
In a split-plot experiment the same levels of B (in the split) are used at each level
of A.
Mix means: B A C
10.29 11.58 13.13
Laboratory x mix means all different in decreasing order: 165, 155, 145 with
B, A, and C.
13.13
Source EMS
Ri + 9GRL + 27cr|
j-j -f 9(7iL + 36O-^
RLij + 9^RL
Mm + ^^RLM + 9(7RM +
RMim + ^^RLM + 9ORM
TMjm + ^^RLM +
RLMijm + ^^RLM
TM,m + ^RLTM + ^^LTM + ^^RTM +
RT Mi,m + ^RLTM + ^^RTM
^2
LT Mj,m ol ^RLTM + ^^LTM
2
RTLMijkm CTE
13.15 a. Split-plot.
Source df EMS
R. r — 1 -|- 15<T^
Mj 2 (^e + ‘^^RM + IQ^A^
RMij 2(r - 1) + ‘^^RM
Tk 4 Gt + 3(TR7’
RT -I- 6^^
2
RT,, 4(r - 1) a; -I- 3(T RT
c • r > 4.
MTj, 8 G; + (j RMT + 2(/) MT]
RMT,j, 8(r - 1) G; G RMT
414 Answers to odd-numbered problems
13.17
Source df EMS
Ri 1 + 18(7^
Mj 2 +
RM,j 2
T, 5 + T
RTik 5
MTj, 10 + ^RMT +
RMT,j, 10 ^6 + ^RMT ^
CHAPTER 14
14.1 Show. SSbiocks ~ 10.125.
14.3 replication I replication II replication III
confound AB confound AC confound BC
14.5 BC.
14.7 SSbiocks ~ 29,153.5 — 884^ + + ^^BCD-
14.9 The principal block contains: (1), ab, ad, bd, c, abc, acd, bed.
14.11 One scheme is to confound ABC, CDE, and ABDE. The principal block contains
(1), ab, de, abde, ace, bee, acd, bed.
Source df
Main effects 5
Two-way interaction 10
Three-way interaction 8
Four-way interaction 4
Five-way interaction 1
Blocks (or ABC, CDE, ABDE) 3
Total 31
14.13 One scheme is to confound ABCD^. The principal block contains 0000, 1110,
1102, 1012, 2111, 2122, 0221, 0210, 0120, 0101, 1211, 1200, 2220, 0022, 0011,
0202, 0112, 1121, 1020, 1001, 1222, 2010, 2201, 2100, 2002, 2021, 2212.
14.15 Verify.
Answers to odd-numbered problems 415
14.17 Confounded ABC, CDE, ADF, BEF, ABDE, BCDF, ACEF gives a principal
block of (1), acd, hce, abf, abde, bcdf, acef, def.
Source df
Total 63
14.19 Yates with seven columns—pool to get temperature and its interactions.
14.21 Confound TABC with blocks (mod 2), L = + Y2 + Y3 + Y4.
14.23 Principal blocks are
Source df
Replication 2
Blocks in replication 9 11
Total 95
CHAPTER 15
15.1 Aliases:
Source df SS
A = C 1 30.25
B = ABC 1 2.25 No tests.
AB = BC 1 0.25
416 Answers to odd-numbered problems
A ^ BC = AB^C^
B = AB^C - AC
C = ABC^ = AB
AB^ = AC^ = BC^
15.9
Source df
Total 63
15.11
Source df SS MS
Totals 23 97.96
Answers to odd-numbered problems 417
Source MS
A 12.44
B 53.40*
AB 0.00
C 33.44
AC 7.06
BC 0.00
D 31.22
AD 42.48*
BD 1.62
CD 0.07
Error 5.73
15.15 a. Replication I: (1), ab, de, abde, bed, acd, bee, ace.
b. BC = DE = ABD = ACE.
c. Replications x three-way = 14 df.
15.17 a. ACD.
Source df MS
Source df
CHAPTER 16
Adjusted
Source df Z xy Z^" Z>'^ df MS
Source df MS
Machines M 1 0.0136
Lumber grade L 1 0.0003
ML 1 0.0066
Replications R 1 0.0001
MR 1 0.0028
LR 1 0.0021
MLR 1 0.0015
Source df SS MS
Source df
Treatments 1
Blocks (men) 6
Error 6
Total 13
}
* • '
I ■ . Index
M , I . . • •
\ , /
- I
I
«
< . •
f S
•4 ‘I
^ •
• *
> \
/ -^1
*♦ ,
r I-,. . .-if*
•> '.^W^va
Index
421
422 Index
I S) , ■^-,1
. •,» \ •.'r ’ .•
s\
A i'#. * -' •
.’.'r
•*, - - ♦
♦ » #
. * *t\’,/'■ V; , 1 1 f'w
r*t \ ■
.■.'Si
< ‘-•'i
.V ^
J
< ■■■ V
.♦ .. . >•'''?■
■” ''\[li .). 'j*. •'■-'i- r ' --•■t.* - ‘ * •^',.-1- V--
' r '
’ '^2^
' : .:^v• - V i,r^‘-
•^-. a ■ \ '4m
■ ^i»r.T
r _J»'W • .*“. • t • • A
.:5.' /' s . ..-
'y'l '•*
■'ii ;■' • . * .'T«f -'^ '-o
-;,'T
■ , = 3»- .' -'*• V.,,
■'’, * ,'•' ‘ *.- ■ •’ Jt-'> ' V*I "i.! » t
•’•: ’ ■•,■» iifj.,1 .;- sJ-f.-'-'r-
•r K'' ■
iT' f ■ . . A
• ■ . .• * '■ ' \ ^ L ’
' V '■^'' t V i- . .* ^*.'V' *2^1
mmi0t
V
■«<
%
I I A
)
I'
IT
i
9 4
ft.
1 ,
■ r“
■'i
.9
4^:
iP-'
«r^ -
V -*«
S>
l‘
' f
<s I