Econometrics - Applied Robust Statistic To Regression Analysis
Econometrics - Applied Robust Statistic To Regression Analysis
David J. Olive
Southern Illinois University
Department of Mathematics
Mailcode 4408
Carbondale, IL 62901-4408
[email protected]
July 6, 2005
Contents
Preface v
1 Introduction 1
1.1 Outlier....s . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.2 Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.3 Complements . . . . . . . . . . . . . . . . . . . . . . . . . . 18
1.4 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
i
3.6 The Double Exponential Distribution . . . . . . . . . . . 77
3.7 The Exponential Distribution . . . . . . . . . . . . . . . . 78
3.8 The Two Parameter Exponential Distribution . . . . . 79
3.9 The Extreme Value Distribution . . . . . . . . . . . . . . 80
3.10 The Gamma Distribution . . . . . . . . . . . . . . . . . . 81
3.11 The Half Normal Distribution . . . . . . . . . . . . . . . 83
3.12 The Logistic Distribution . . . . . . . . . . . . . . . . . . 84
3.13 The Lognormal Distribution . . . . . . . . . . . . . . . . . 84
3.14 The Normal Distribution . . . . . . . . . . . . . . . . . . . 85
3.15 The Pareto Distribution . . . . . . . . . . . . . . . . . . . 87
3.16 The Poisson Distribution . . . . . . . . . . . . . . . . . . . 88
3.17 The Power Distribution . . . . . . . . . . . . . . . . . . . . 88
3.18 The Rayleigh Distribution . . . . . . . . . . . . . . . . . . 89
3.19 The Student’s t Distribution . . . . . . . . . . . . . . . . 89
3.20 The Truncated Extreme Value Distribution . . . . . . . 90
3.21 The Uniform Distribution . . . . . . . . . . . . . . . . . . 91
3.22 The Weibull Distribution . . . . . . . . . . . . . . . . . . . 91
3.23 Complements . . . . . . . . . . . . . . . . . . . . . . . . . . 92
3.24 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93
ii
6 Regression Diagnostics 185
6.1 Numerical Diagnostics . . . . . . . . . . . . . . . . . . . . 185
6.2 Graphical Diagnostics . . . . . . . . . . . . . . . . . . . . . 188
6.3 Outlier Detection . . . . . . . . . . . . . . . . . . . . . . . 192
6.4 A Simple Plot for Model Assessment . . . . . . . . . . . 195
6.5 Complements . . . . . . . . . . . . . . . . . . . . . . . . . . 202
6.6 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 203
iii
10.5 Breakdown . . . . . . . . . . . . . . . . . . . . . . . . . . . . 295
10.6 Algorithms for the MCD Estimator . . . . . . . . . . . . 296
10.7 Theory for CMCD Estimators . . . . . . . . . . . . . . . 298
10.8 Complements . . . . . . . . . . . . . . . . . . . . . . . . . . 310
10.9 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 312
12 1D Regression 337
12.1 Estimating the Sufficient Predictor . . . . . . . . . . . . 340
12.2 Visualizing 1D Regression . . . . . . . . . . . . . . . . . . 346
12.3 Predictor Transformations . . . . . . . . . . . . . . . . . . 358
12.4 Variable Selection . . . . . . . . . . . . . . . . . . . . . . . 359
12.5 Inference . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 369
12.6 Complements . . . . . . . . . . . . . . . . . . . . . . . . . . 372
12.7 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 375
iv
Preface
v
Preface vi
estimator are derived, but the examples and software use an inconsistent
“perfect classification” procedure. In this text, some practical estimators
that have good statistical properties are developed (see Theorems 8.7, 10.14
and 10.15), and some effort has been made to state whether the “perfect
classification” or “asymptotic” paradigm is being used.
The majority of the statistical procedures described in Hampel, Ronchetti,
Rousseeuw and Stahel (1986), Huber (1981), and Rousseeuw and Leroy
(1987) assume that outliers are present or that the true underlying error
distribution has heavier tails than the assumed model. However, these three
references and some of the papers in Stahel and Weisberg (1991a,b) and
Maddela and Rao (1997) do discuss other departures from the assumed
model. Other texts on distributional robustness include Atkinson and Riani
(2000), Atkinson, Riani and Cerioli (2004), Dell’Aquila and Ronchetti (2005),
Hettmansperger and McKean (1998), Hoaglin, Mosteller and Tukey (1983),
Insightful (2002), Jureckova and Sen (1996), Marazzi (1993), Maronna (2006),
Morgenthaler, Ronchetti, and Stahel (1993), Morgenthaler and Tukey (1991),
Müller (1997), Rey (1978), Rieder (1996), Shevlyakov and Vilchevski (2002),
Staudte and Sheather (1990) and Wilcox (2005). Diagnostics and outliers
are discussed in Atkinson (1985), Barnett and Lewis (1994), Belsley, Kuh,
and Welsch (1980), Chatterjee and Hadi (1988), Cook and Weisberg (1982),
Fox (1991), Hawkins (1980) and Iglewicz and Hoaglin (1993).
Several textbooks on statistical analysis and theory also discuss robust
methods. For example, see Dodge and Jureckova (2000), Gentle (2002),
Gnanadesikan (1997), Hamilton (1992), Seber and Lee (2003), Thode (2002)
and Wilcox (2001, 2003).
Besides distributional robustness, this book also considers regression graph-
ics procedures that are useful even when the 1D regression model is unknown
or misspecified. 1D regression and regression graphics procedures are de-
scribed in Cook and Weisberg (1999a), Cook (1998a) and Li (2000).
A unique feature of this text is the discussion of the interrelationships be-
tween distributionally robust procedures and regression graphics with focus
on 1D regression. A key assumption for regression graphics is that the predic-
tor distribution is approximately elliptically contoured. Ellipsoidal trimming
(based on robust estimators of multivariate location and dispersion) can be
used to induce this condition. An important regression graphics technique is
dimension reduction: assume that there are p predictors collected in a p × 1
vector x. Then attempt to reduce the dimension of the predictors from p to
Preface viii
Background
This course assumes that the student has had considerable exposure to
statistics, but is at a much lower level than most texts on distributionally
robust statistics. Calculus and a course in linear algebra are essential. Fa-
miliarity with least squares regression is also assumed and could come from
econometrics or numerical linear algebra, eg Weisberg (2005), Datta (1995),
Golub and Van Loan (1989) or Judge, Griffiths, Hill, Lütkepohl and Lee
(1985). The matrix representation of the multiple linear regression model
should be familiar. An advanced course in statistical inference, especially
one that covered convergence in probability and distribution, is needed for
several sections of the text. Casella and Berger (2002), Poor (1988) and
White (1984) easily meet this requirement.
There are other courses that would be useful but are not required. An
advanced course in least squares theory or linear models can be met by Seber
and Lee (2003) in statistics, White (1984) in economics, and Porat (1993) in
electrical engineering. Knowledge of the multivariate normal distribution at
the level of Johnson and Wichern (1988) would be useful. A course in pattern
recognition, eg Duda, Hart and Stork (2000), also covers the multivariate
normal distribution.
If the students have had only one calculus based course in statistics (eg
DeGroot 1975 or Wackerly, Mendenhall and Scheaffer 2002), then cover Ch.
1, 2.1–2.5, 4.6, Ch. 5, Ch. 6, 7.6, part of 8.2, 9.2, 10.1, 10.2, 10.3, 10.6,
10.7, 11.1, 11.3, Ch. 12 and Ch. 13. (This will cover the most important
material in the text. Many of the remaining sections are for Ph.D. students
and experts in robust statistics.) Many of the Chapter 5 homework problems
were used in the author’s multiple linear regression course, and many Chapter
13 problems were used in the author’s categorical data analysis course.
Some of the applications in this text include
• Robust parameter estimation using the sample median and the sample
median absolute deviation is described on p. 35–37 and in Chapter 3.
• Section 6.3 shows how to use the forward response plot to detect outliers
and to assess the adequacy of the multiple linear regression model.
• Section 6.4 shows how to use the FY plot to detect outliers and to
assess the adequacy of very general regression models of the form y =
m(x) + e.
• Section 7.6 provides the resistant mbareg estimator for multiple linear
regression which is useful for teaching purposes.
• Section 8.2 shows how to modify the inconsistent zero breakdown esti-
mators for LMS and LTS (such √ as lmsreg) so that the resulting modifi-
cation is an easily computed n consistent high breakdown estimator.
√
• Sections 10.6 and 10.7 provide the easily computed robust n consis-
tent HB covmba estimator for multivariate location and dispersion. It
is also shown how to modify the inconsistent zero breakdown cov.mcd
√
estimator so that the resulting modification is an easily computed n
consistent high breakdown estimator. Application are numerous.
• Section 11.1 shows that the DD plot can be used to detect multivariate
outliers and as a diagnostic for whether the data is multivariate nor-
mal or from some other elliptically contoured distribution with second
moments.
Preface xi
• Section 11.3 suggests the resistant tvreg estimator for multiple linear
regression that can be modified to create a resistant weighted MLR
estimator if the weights wi are known.
> source("A:/rpack.txt")
> source("A:/rpack.txt")
will enter the functions into Splus. Creating a special workspace for the
functions may be useful.
Type ls(). About 40 R/Splus functions from rpack.txt should appear. In
R, enter the command q(). A window asking “Save workspace image?” will
appear. Click on No to remove the functions from the computer (clicking on
Yes saves the functions on R, but you have the functions on your disk).
Similarly, to download the text’s R/Splus data sets, save robdata.txt on a
disk and use the following command.
> source("A:/robdata.txt")
MCD, MVE, projection, repeated median and S estimators. Two stage esti-
mators that need an initial high breakdown estimator from the above list are
even less practical to compute. These estimators include the cross checking,
MM, one step GM, one step GR, REWLS, tau and t type estimators. Also,
although two stage estimators tend to inherit the breakdown value of the
initial estimator, their outlier resistance as measured by maximal bias tends
to decrease sharply. Typically the implementations for these estimators are
not given, impractical to compute, or result in a zero breakdown estimator
that is often inconsistent. The inconsistent zero breakdown implementations
and ad hoc procedures should usually only be used as diagnostics for outliers
and other model misspecifications, not for inference.
Many of the ideas in the HB literature are good, but the ideas were
premature for applications without a computational and theoretical break-
through. This text, Olive(2004a) and Olive and Hawkins (2006) provide
this breakthrough and show that simple modifications to elemental basic√ re-
sampling or concentration algorithms result in the easily computed HB n
consistent CMCD estimator for multivariate location and dispersion (MLD)
and CLTS estimator for multiple linear regression (MLR). The Olive (2004a)
MBA estimator is a special case of the CMCD estimator and is much faster
than the inconsistent zero breakdown Rousseeuw and Van Driessen (1999)
FMCD estimator. The Olive (2005) resistant MLR estimators also have good
statistical properties. See Sections 7.6, 8.2, 10.7, 11.4, Olive (2004a, 2005),
Hawkins and Olive (2002) and Olive and Hawkins (2006).
As an illustration for how the CMCD estimator improves the ideas from
the HB literature, consider the He and Wang (1996) cross checking estima-
tor that uses the classical estimator if it is close to the robust estimator,
and uses the robust estimator otherwise. The resulting estimator is an HB
asymptotically efficient estimator if a consistent HB robust estimator is used.
He and Wang (1997) show that the all elemental subset approximation to S
estimators is a consistent HB MLD estimator that could be used in the cross
checking estimator, but then the resulting cross checking estimator is im-
practical to compute. If the (basic resampling MVE or) FMCD estimator
is used, then the cross checking estimator is practical to compute but has
zero breakdown since the FMCD and classical estimators both have zero
breakdown. Since the FMCD estimator is inconsistent and highly variable,
the probability that the FMCD estimator and classical estimator are close
does not go to one as n →√∞. Hence the cross checking estimator is also
inconsistent. Using the HB n consistent CMCD estimator results in an HB
Preface xiv
Introduction
Y x|β T x. (1.1)
1
CHAPTER 1. INTRODUCTION 2
Yi = g(xTi β, ei ) (1.2)
Y = xT β + e (1.4)
t(y) = β T x + e. (1.6)
exp(β T x)
ρ(β T x) = .
1 + exp(β T x)
CHAPTER 1. INTRODUCTION 3
ri = Yi − m(xTi β̂)
Z = t(Y ) = xT β + e (1.9)
1.1 Outlier....s
The main message of this book is that robust regression is extremely useful
in identifying outliers ....
Rousseeuw and Leroy (1987, p. vii)
Hampel, Ronchetti, Rousseeuw and Stahel (1986, p. 36) state that the first
and most important step in robustification is the rejection of distant outliers.
In the literature there are two important paradigms for robust procedures.
The perfect classification paradigm considers a fixed data set of n cases of
which 0 ≤ d < n/2 are outliers. The key assumption for this paradigm is
that the robust procedure perfectly classifies the cases into outlying and non-
outlying (or “clean”) cases. The outliers should never be blindly discarded.
Often the clean data and the outliers are analyzed separately.
The asymptotic paradigm uses an asymptotic distribution to approximate
the distribution of the estimator when the sample size n is large. An impor-
tant example is the central limit theorem (CLT): let Y1 , ..., Yn be iid with
mean µ and standard deviation σ; ie, the Yi ’s follow the location model
Y = µ + e.
Then
√ 1 n
Yi − µ) → N(0, σ 2 ).
D
n(
n i=1
1.2 Applications
One of the key ideas of this book is that the data should be examined with
several estimators. Often there are many procedures that will perform well
when the model assumptions hold, but no single method can dominate every
CHAPTER 1. INTRODUCTION 6
other method for every type of model violation. For example, OLS is best
for multiple linear regression when the iid errors are normal (Gaussian) while
L1 is best if the errors are double exponential. Resistant estimators may
outperform classical estimators when outliers are present but be far worse if
no outliers are present.
Portnoy and Mizera (1999) note that different multiple linear regression
estimators tend to estimate β in the iid constant variance symmetric error
model, but otherwise each estimator estimates a different parameter. Hence
a plot of the residuals or fits from different estimators should be useful for
detecting departures from this very important model. The “RR plot” is a
scatterplot matrix of the residuals from several regression fits. Tukey (1991)
notes that such a plot will be linear with slope one if the model assumptions
hold. Let the ith residual from the jth fit β̂ j be ri,j = Yi − xTi β̂ j where
the superscript T denotes the transpose of the vector and (Yi , xTi ) is the ith
observation. Then
ri,1 − ri,2 = xTi (β̂1 − β̂2 )
≤ xi (β̂1 − β + β̂ 2 − β).
The RR plot is simple to use since if β̂ 1 and β̂ 2 have good convergence
rates and if the predictors xi are bounded, then the residuals will cluster
tightly about the identity line (the unit slope line through the origin) as n
increases to ∞. For example, plot the least squares residuals versus the L1
residuals. Since OLS and L1 are consistent, the plot should be linear with
slope one when the regression assumptions hold, but the plot should not have
slope one if there are Y –outliers since L1 resists these outliers while OLS does
not. Making a scatterplot matrix of the residuals from OLS, L1 , and several
other estimators can be very informative.
Example 1.1. Gladstone (1905–1906) attempts to estimate the weight
of the human brain (measured in grams after the death of the subject) using
simple linear regression with a variety of predictors including age in years,
height in inches, head height in mm, head length in mm, head breadth in mm,
head circumference in mm, and cephalic index (divide the breadth of the head
by its length and multiply by 100). The sex (coded as 0 for females and 1
for males) of each subject was also included. The variable cause was coded
as 1 if the cause of death was acute, as 3 if the cause of death was chronic,
and coded as 2 otherwise. A variable ageclass was coded as 0 if the age was
under 20, as 1 if the age was between 20 and 45, and as 3 if the age was
CHAPTER 1. INTRODUCTION 7
• • •
••• • ••• • • •• •
•• ••
• •
• •
•••••• •••• • •• •••••••
• •••• • •
•
••• • • •••• •
• ••
••
•• • •• •• • •
• • ••••••••• •
•• • • • •
•••••••••• ••••••••••• • •• ••••••••••••••••••••••
•••••••••••••••
• ••••••••••••••• •
• •• • •••
• •
•• • •
• • •
••••••• • •• ••••••••••
• •
•••
••
•
••••
••••• • ••
••••••••• • • • • ••••••••••••••••••••••
•••••••••••• ••••••• • • •••••••••••••••••• ••
•••••••••••• • ••••••••• • • •• • • • •
• •• ••••• • •
•• ••••
• • • ••
• • •
• ••• ••
•• • • ••••• •
• •••• • ••••••• • • •
•• • • • •
• • ••• • • • ••••• • •
•• •• • • •••••••••••••••••••
•••
•••••••••• • • •••••••••••••
••••••• • •
• ••••••••• • •
••••• •
••••••••••••••• •• ••••••••••••••••••••
•••••••••••••••••• •
•••
••••• • •
• •••••••••••••••••••••••••••••
• •
• ••••••••••••• ••••
••••••••••••• • •••••••••• • •
••••••••••••
• • • • •••••••••••••••• •
•• • • •• •••
•••• • ••••••
• • •
• • •
• • •• • •
• • •
• • •
• • •
• • •• • •
• ••
• •• • • •• • • •• • • • • •
•••• ••••••••••••••••••••••••••••••••••••• •••••• • ••••••••••••••••••• •••••••• •
• •••••••••• ••••••••••••••••••••• ••••••• •
• ••••••••••••••••••
•• •• ••••••••
• •••••••••••••••••• • •••••••••••••••••••••••••••••••••••••••••••••••••
• •••••••••••••• • ••• • •• •••••••••••• •• •
• • •
• • •
• • • • ••• ••• • •
••••• • • •••••
• •• • •• ••••••
•• • • •
•
•••••••••••••••••••••••• •• •••• •• •••
••••••••••••• • • ••
••••••••••••
• • • • • •• • • •• ••••• •• •
• • •••••••••••••••••••••••••••••••••••••• • •••••••••••••••••••••••• •••••••••••••••• • • ••
• •• • ••• •• • ••••••••••••••••••••••••• • •••••••• •
••••••••••••••••••••••••• •••••••••• •••••••••••
•••••••••••••••• • • •••••••••••• •••••••••••
•• •• • •
• •• • ••• •
•
• • • • ••
• • • • • • •••• •
• • • •••••••••••••••
••• •••••••••••••
•••••••••
• • • • •••••••••••••••••••••••••••••••••••••
• • •
• • •• • • • •
••
• •• • •• • • •• •
• •
• •
• • • • •
• • • ••• • • •• • •
• • • •• •• • •
• •• ••• •
• • • • • • • •• • ••• ••• • • •
• • •••• ••• • ••••• • • • • • ••• •• • •
• • ••• • • • • • • • • • •••• •• ••• • •
• • • • • • ••• ••• •••• •••• •• • •• ••• • •
• ••• ••••• ••• • • • •••• ••••••• • ••• •
• •• ••••• ••
••• ••• ••••••••• •• •• • ••• ••
• • • • • •• ••• • •
• • • • •••• •••••• • • •• •• • ••••• ••••• ••
• •• • • •
• • • •• • • • • •• • • ••••••••••••••••• • •
• •• • ••••••••••••••• •• • • • • • •••• • ••• • •
• • •• •••• •••• ••••• • •• • • • ••• • •
•• • • • •• • • •• • •• •
•• •• •• • •• •••• • •
••• • •••• •••• •• •
• • • • •• • • •• • •• • • •
• • •• • •• • ••••• • •
• • • •• • •
•• • • •• • • •
••• •••
• •
• • •
• • • •
• • • • ••••••••••
• •• • •
• •• • • • ••• •••••••••••••••••••••••
•••••••••••••
• • • • •••••••••••••••••••••••••••••••••• •
• •• ••• • • • •• •• •• • • •
• ••
• • • • •
• • •• ••••••••••• •••••••••••••••••••••••••••• • •••• ••• • • • • • • •
• •• • • • •••••••••••••• • ••• •••• •• ••• • • • • •
• • • • •• •• •
•••• ••••• • • •••••• •••••• ••••• •••• ••• • • • •• •
• •• • • •• •• •• •• • • • • • •
•• •• •
••• •• • • • •
• •• • •
• ••• •• • •• • •• ••
•
• •
• ••••••••••• ••••••••••
•••••••••••
•••••••••
•••••
•••••••••••••••••••••••
•••••••
••••• ••••
•
••
• • •
over 45. Head size is the product of the head length, head breadth, and head
height.
The data set contains 276 cases, and we decided to use multiple linear
regression to predict brain weight using the six head measurements height,
length, breadth, size, cephalic index and circumference as predictors. Cases
188 and 239 were deleted because of missing values. There are five infants
(cases 238, 263-266) of age less than 7 months that are x-outliers. Nine
toddlers were between 7 months and 3.5 years of age, four of whom appear
to be x-outliers (cases 241, 243, 267, and 269). Figure 1.1 shows an RR
plot comparing the OLS, L1 , ALMS and ALTS fits. ALMS is the default
version of the R/Splus function lmsreg while ALTS is the default version of
ltsreg, and these two resistant estimators are described further in Chapter
7. Attempts to reproduce Figure 1.1 (made in 1997 with an old version of
Splus) will fail due to changes in the R/Splus code. Note that Figure 1.1
suggests that three of the methods are producing approximately the same
fits while the ALMS estimator is fitting 9 of the 274 points in a different
manner. These 9 points correspond to five infants and four toddlers that are
x-outliers.
An obvious application of outlier resistant methods is the detection of
outliers. Generally robust and resistant methods can only detect certain
configurations of outliers, and the ability to detect outliers rapidly decreases
as the sample size n and the number of predictors p increase. When the
Gladstone data was first entered into the computer, the variable head length
was inadvertently entered as 109 instead of 199 for case 119. Residual plots
for six Splus regression estimators (described further in Section 7.2, KLMS
and KLTS used options that should generally detect more outliers than the
default versions of lmsreg and ltsreg) are shown in Figure 1.2. In 1997,
ALMS and the classical OLS and L1 estimators failed to identify observation
119 as unusual. Eventually this coding error was detected and corrected.
Example 1.2. Buxton (1920, p. 232-5) gives 20 measurements of 88
men. Height was the response variable while an intercept, head length, nasal
height, bigonal breadth, and cephalic index were used as predictors in the
multiple linear regression model. Observation 9 was deleted since it had
missing values. Five individuals, numbers 62–66, were reported to be about
0.75 inches tall with head lengths well over five feet! Figure 1.3 shows that
the outliers were accommodated by all of the Splus estimators, except KLMS.
The Buxton data is also used to illustrate robust multivariate location and
CHAPTER 1. INTRODUCTION 10
• • •
••
• ••
•
•• • ••
• •• •
• •• • • ••
•• • •• •• •
••• • ••
• •••••
•••••• • • • •• ••
••• • ••••••• •• •• ••••••••
•
• ••• ••• • •• •• •
•• •• •••••• • •••• • •
• ••
•• • • •••••••• • • •••
••• ••
• ••••• • •• •••
••••••• • • •
• •••• • • ••
• ••••
•• •••• • •••• •
• •
••• • •
•• •• •
• •
••• •• • •
• ••••
•• ••• •
•• • •
•
• ••
• • •
• ••• ••
• ••••••••••••••••••••• •
•••
•
• • ••
•
•• • ••
• ••• •
• •• •
• ••
• •• •••••••
•• • • ••••••••
• •••••• • •••••• ••••
• • ••• •
•••
••• •••• ••
•• • •••
• ••••••• •• •••
• • ••••
•• • •
• • ••••• • ••••
••
•• • •
•
• ••• •
•••• •
••••• •
•
•
•
•
•
• • •
Figure 1.3: Buxton data, the outliers do not have large residuals.
CHAPTER 1. INTRODUCTION 11
should be used. The cleaned data will be used to show that the bulk of the
data is well approximated by the statistical model, but the full data set will
be used along with the cleaned data for prediction and for description of the
entire population.
To illustrate the above discussion, consider the multiple linear regression
model
Y = Xβ + e (1.10)
where Y is an n × 1 vector of dependent variables, X is an n × p matrix
of predictors, β is a p × 1 vector of unknown coefficients, and e is an n × 1
vector of errors. The ith case (Yi , xTi ) corresponds to the ith row xTi of X
and the ith element Yi of Y . Assume that the errors ei are iid zero mean
normal random variables with variance σ 2 .
Finding prediction intervals for future observations is a standard problem
in regression. Let β̂ denote the ordinary least squares (OLS) estimator of β
and let n 2
r
MSE = i=1 i
n−p
where ri = Yi − xTi β̂ is the ith residual. Following Neter, Wasserman, Nacht-
sheim and Kutner (1996, p. 235), a (1 − α)100% prediction interval (PI) for
a new observation Yh corresponding to a vector of predictors xh is given by
a) No Outliers b) No Outliers
40
0.04
30
0.0
R
Y
20
-0.04
10
-0.08
0
0 10 20 30 40 0 10 20 30 40
FIT FIT
96
15
40
10
169
30
182 76
5
R
Y
96 76 200
20
0 6
10
169
-5
182
48 48
0
0 10 20 30 40 0 10 20 30 40
FIT FIT
Assume that the data have been perfectly classified into nc clean cases and
no outlying cases where nc + no = n. Also assume that no outlying cases will
fall within the PI. Then the PI is valid if Yh is clean, and
where Ŷh and se(pred) are obtained after performing OLS on the nc clean
cases. For example, if α = 0.1 and γ = 0.08, then 1 − α∗ ≈ 0.98. Since γ will
be estimated from the data, the coverage will only be approximately valid.
The following example illustrates the procedure.
Example 1.4. STATLIB provides a data set (see Johnson 1996) that is
available from the website (http://lib.stat.cmu.edu/datasets/bodyfat). The
CHAPTER 1. INTRODUCTION 14
data set includes 252 cases, 14 predictor variables, and a response variable
Y = bodyfat. The correlation between Y and the first predictor x1 = density
is extremely high, and the plot of x1 versus Y looks like a straight line except
for four points. If simple linear regression is used, the residual plot of the
fitted values versus the residuals is curved and five outliers are apparent.
The curvature suggests that x21 should be added to the model, but the least
squares fit does not resist outliers well. If the five outlying cases are deleted,
four more outliers show up in the plot. The residual plot for the quadratic fit
looks reasonable after deleting cases 6, 48, 71, 76, 96, 139, 169, 182 and 200.
Cases 71 and 139 were much less discrepant than the other seven outliers.
These nine cases appear to be outlying at random: if the purpose of the
analysis was description, we could say that a quadratic fits 96% of the cases
well, but 4% of the cases are not fit especially well. If the purpose of the
analysis was prediction, deleting the outliers and then using the clean data to
find a 99% prediction interval (PI) would not make sense if 4% of future cases
are outliers. To create a nominal 90% PI for future cases from this population,
make a classical 100(1−α∗ ) PI from the clean cases where 1−α∗ = 0.9/(1−γ).
For the bodyfat data, we can take 1−γ ≈ 1−9/252 ≈ 0.964 and 1−α∗ ≈ 0.94.
Notice that (0.94)(0.96) ≈ 0.9.
Figure 1.4 is useful for presenting the analysis. The top two plots have
the nine outliers deleted. Figure 1.4a is a forward response plot of the fitted
values Ŷi versus the response Yi while Figure 1.4b is a residual plot of the
fitted values Ŷi versus the residuals ri . These two plots suggest that the
multiple linear regression model fits the bulk of the data well. Next consider
using weighted least squares where cases 6, 48, 71, 76, 96, 139, 169, 182 and
200 are given weight zero and the remaining cases weight one. Figure 1.4c
and 1.4d give the forward response plot and residual plot for the entire data
set. Notice that seven of the nine outlying cases can be seen in these plots.
The classical 90% PI using x = (1, 1, 1)T and all 252 cases was Ŷh ±
t0.95,249se(pred) = 46.3152 ± 1.651(1.3295) = (44.12, 48.51). When the 9 out-
liers are deleted, nc = 243 cases remain. Hence the 90% PI using Equation
(1.13) with 9 cases deleted was Ŷh ±t0.97,240se(pred) = 44.961±1.88972(0.0371)
= (44.89, 45.03). The classical PI is about 31 times longer than the new PI.
For the next application, consider a response transformation model
y = t−1
λo (x β + e)
T
CHAPTER 1. INTRODUCTION 15
tλo (y) = xT β + e
follows a multiple linear regression (MLR) model where the response variable
yi > 0 and the power transformation family
yλ − 1
tλ(y) ≡ y (λ) = (1.14)
λ
for λ = 0 and y (0) = log(y).
The following simple graphical method for selecting response transforma-
tions can be used with any good classical, robust or Bayesian MLR estimator.
Let zi = tλ(yi ) for λ = 1, and let zi = yi if λ = 1. Next, perform the multiple
linear regression of zi on xi and make the forward response plot of zˆi versus
zi . If the plotted points follow the identity line, then take λo = λ. One plot
is made for each of the eleven values of λ ∈ Λ, and if more than one value of
λ works, take the simpler transformation or the transformation that makes
the most sense to subject matter experts. (Note that this procedure can be
modified to create a graphical diagnostic for a numerical estimator λ̂ of λo
by adding λ̂ to Λ.) The following example illustrates the procedure.
Example 1.5. Box and Cox (1964) present a textile data set where
samples of worsted yarn with different levels of the three factors were given
a cyclic load until the sample failed. The goal was to understand how y =
the number of cycles to failure was related to the predictor variables. Figure
1.5 shows the forward response plots for two MLR estimators: OLS and
the R/Splus function lmsreg. Figures 1.5a and 1.5b show that a response
transformation is needed while 1.5c and 1.5d both suggest that log(y) is the
appropriate response transformation. Using OLS and a resistant estimator
as in Figure 1.5 may be very useful if outliers are present.
The textile data set is used to illustrate another graphical method for
selecting the response transformation tλ in Section 5.1.
Another important application is variable selection: the search for a sub-
set of predictor variables that can be deleted from the model without impor-
tant loss of information. Section 5.2 gives a graphical method for assessing
variable selection for multiple linear regression models while Section 12.4
gives a similar method for 1D regression models.
CHAPTER 1. INTRODUCTION 16
3000
3000
2000
2000
Y
Y
1000
1000
0
0
-500 0 500 1000 1500 2000 0 200 400 600 800
FIT RFIT
8
7
7
LOGY
LOGY
6
6
5
5 6 7 8 5 6 7 8
WFIT RWFIT
Figure 1.5: OLS and LMSREG Suggest Using log(y) for the Textile Data
The basic idea is to obtain fitted values from the full model and the
candidate submodel. If the candidate model is good, then the plotted points
in a plot of the submodel fitted values versus the full model fitted values
should follow the identity line. In addition, a similar plot should be made
using the residuals.
A problem with this idea is how to select the candidate submodel from
the nearly 2p potential submodels. One possibility would be to try to order
the predictors in importance, say x1, ..., xp. Then let the kth model contain
the predictors x1 , x2, ..., xk for k = 1, ..., p. If the predicted values from the
submodel are highly correlated with the predicted values from the full model,
then the submodel is “good.” This idea is useful even for extremely compli-
cated models. Section 12.4 will show that the all subsets, forward selection
and backward elimination techniques of variable selection for multiple lin-
ear regression will often work for the 1D regression model provided that the
Mallows’ Cp criterion is used.
Example 1.6. The Boston housing data of Harrison and Rubinfeld
(1978) contains 14 variables and 506 cases. Suppose that the interest is
in predicting the per capita crime rate from the other variables. Variable
CHAPTER 1. INTRODUCTION 17
OLS View
500
Y
0
-500
X %*% bols
selection for this data set is discussed in much more detail in Section 12.4.
Another important topic is fitting 1D regression models given by Equation
(1.2) where g and β are both unknown. Many types of plots will be used
in this text and a plot of x versus y will have x on the horizontal axis and
y on the vertical axis. This notation is also used by the software packages
Splus (MathSoft 1999ab) and R, the free version of Splus available from
(http://www.r-project.org/). The R/Splus commands
X <- matrix(rnorm(300),nrow=100,ncol=3)
Y <- (X %*% 1:3)^3 + rnorm(100)
were used to generate 100 trivariate Gaussian predictors x and the response
Y = (βT x)3 + e where e ∼ N(0, 1). This is a model of form (1.3) where m is
the cubic function.
An amazing result is that the unknown function m can often be visualized
by the “OLS view,” a plot of the OLS fit (possibly ignoring the constant)
versus Y generated by the following commands.
bols <- lsfit(X,Y)$coef[-1]
plot(X %*% bols, Y)
CHAPTER 1. INTRODUCTION 18
The OLS view, shown in Figure 1.6, can be used to visualize m and
for prediction. Note that Y appears to be a cubic function of the OLS fit
and that if the OLS fit = 0, then the graph suggests using Ŷ = 0 as the
predicted value for Y . This plot and modifications will be discussed in detail
in Chapters 12 and 13.
This section has given a brief outlook of the book. Also look at the preface
and table of contents, and then thumb through the remaining chapters to
examine the procedures and graphs that will be developed.
1.3 Complements
Many texts simply present statistical models without discussing the process
of model building. An excellent paper on statistical models is Box (1979).
The concept of outliers is rather vague although Barnett and Lewis (1994),
Davies and Gather (1993) and Gather and Becker (1997) give outlier models.
Also see Beckman and Cook (1983) for history.
Outlier rejection is a subjective or objective method for deleting or chang-
ing observations which lie far away from the bulk of the data. The modified
data is often called the “cleaned data.” See Rousseeuw and Leroy (1987,
p. 106, 161, 254, and 270), Huber (1981, p. 4-5, and 19), and Hampel,
Ronchetti, Rousseeuw and Stahel (1986, p. 24, 26, and 31). Data editing,
screening, truncation, censoring, Winsorizing, and trimming are all methods
for data cleaning. David (1981, ch. 8) surveys outlier rules before 1974, and
Hampel, Ronchetti, Rousseeuw and Stahel (1986, Section 1.4) surveys some
robust outlier rejection rules. Outlier rejection rules are also discussed in
Hampel (1985), Simonoff (1987a,b), and Stigler (1973b).
Robust estimators can be obtained by applying classical methods to the
cleaned data. Huber (1981, p. 4-5, 19) suggests that the performance of such
methods may be more difficult to work out than that of robust estimators
such as the M-estimators, but gives a procedure for cleaning regression data.
Staudte and Sheather (1990, p. 29, 136) state that rejection rules are the least
understood and point out that for subjective rules where the cleaned data is
assumed to be iid, one can not find an unconditional standard error estimate.
Even if the data consists of observations which are iid plus outliers, some
“good” observations will usually be deleted while some “bad” observations
will be kept. In other words, the assumption of perfect classification is often
CHAPTER 1. INTRODUCTION 19
unreasonable.
The graphical method for response transformations illustrated in Example
1.5 was suggested by Olive (2004b).
Seven important papers that influenced this book are Hampel (1975),
Siegel (1982), Devlin, Gnanadesikan and Kettenring (1981), Rousseeuw (1984),
Li and Duan (1989), Cook and Nachtsheim (1994) and Rousseeuw and Van
Driessen (1999). The importance of these papers will become clearer later in
the text.
An excellent text on regression (using 1D regression models such as (1.1))
is Cook and Weisberg (1999a). A more advanced text is Cook (1998a). Also
see Cook (2003), Horowitz (1998), Li (2000) and Weisberg (2005).
This text will use the software packages Splus (MathSoft (now Insightful)
1999ab) and R, a free version of Splus available from the website (http://www.
r-project.org/), and Arc (Cook and Weisberg 1999a), a free package available
from the website (http://www.stat.umn.edu/arc).
Section 14.2 of this text, Becker, Chambers, and Wilks (1988), and Ven-
ables and Ripley (1997) are useful for R/Splus users. The websites (http://
www.burns-stat.com/), (http://lib.stat.cmu.edu/S/splusnotes) and (http://
www.isds.duke.edu/computing/S/Snotes/Splus.html) also have useful infor-
mation.
The Gladstone, Buxton, bodyfat and Boston housing data sets are avail-
able from the text’s website under the file names gladstone.lsp, buxton.lsp,
bodfat.lsp and boston2.lsp.
1.4 Problems
PROBLEMS WITH AN ASTERISK * ARE ESPECIALLY USE-
FUL.
1.1∗. Using the notation on p. 6, let Ŷi,j = xTi β̂j and show that
ri,1 − ri,2 = Ŷi,1 − Ŷi,2 .
R/Splus Problems
1.2∗. a) Using the R/Splus commands on p. 17, reproduce a plot like
Figure 1.6. Once you have the plot you can print it out directly, but it will
CHAPTER 1. INTRODUCTION 20
d) Include the plot in Word using commands similar to those given in b).
e) Do the two plots look similar? Can you see the cubic function?
1.3∗. a) Enter the following R/Splus function that is used to illustrate
the central limit theorem when the data Y1 , ..., Yn are iid from an exponential
distribution. The function generates a data set of size n and computes Y 1
from the data set. This step is repeated nruns = 100 times. The output is
a vector (Y 1, Y 2 , ..., Y 100 ). A histogram of these means should resemble a
symmetric normal density once n is large enough.
of.” A window will appear. Double click on L1: Fit–Values and then double
click on L1:Residuals. Then L1: Fit–Values should appear in the H box and
L1:Residuals should appear in the V box. Click on OK to obtain the plot.
d) The graph can be printed with the menu commands “File>Print,” but
it will generally save paper by placing the plots in the Word editor. Activate
Word (often by double clicking on a Word icon). Click on the screen and
type “Problem 1.4.” In Arc, use the menu command “Edit>Copy.” In Word,
use the menu commands “Edit>Paste.”
e) In your Word document, write “1.4e)” and state whether the points
cluster about the horizontal axis with no pattern. If curvature is present,
then the multiple linear regression model is not appropriate.
f) After editing your Word document, get a printout by clicking on the
printer icon or by using the menu commands “File>Print.” To save your
output on your diskette, use the Word menu commands “File > Save as.” In
the Save in box select “3 1/2 Floppy(A:)” and in the File name box enter
HW1d4.doc. To exit from Word and Arc, click on the “X” in the upper right
corner of the screen. In Word a screen will appear and ask whether you want
to save changes made in your document. Click on No. In Arc, click on OK.
Warning: The following problem uses data from the book’s web-
page. Save the data files on a disk. Next, get in Arc and use the menu
commands “File > Load” and a window with a Look in box will appear.
Click on the black triangle and then on 3 1/2 Floppy(A:). Then click twice
on the data set name, eg, bodfat.lsp. These menu commands will be de-
noted by “File > Load > 3 1/2 Floppy(A:) > bodfat.lsp” where the data file
(bodfat.lsp) will depend on the problem.
If the free statistics package Arc is on your personal computer (PC),
there will be a folder Arc with a subfolder Data that contains a subfolder
Arcg. Your instructor may have added a new folder mdata in the subfolder
Data and added bodfat.lsp to the folder mdata. In this case the Arc menu
commands “File > Load > Data > mdata > bodfat.lsp” can be used.
1.5∗. This text’s webpage has several files that can be used by Arc.
Chapter 14 explains how to create such files.
a) Use the Arc menu commands “File > Load > 3 1/2 Floppy(A:) >
bodfat.lsp” to activate the file bodfat.lsp.
CHAPTER 1. INTRODUCTION 23
dow, and click on the Full quad. circle. Then click on OK. These commands
will fit the quadratic model y = x1 + x12 + e without using the deleted cases.
Make a residual plot of L4:Fit-Values versus L4:Residuals and a forward re-
sponse plot of L4:Fit-Values versus y. For both plots place the fitted values
in the H box and the other variable in the V box. Include these two plots in
Word.
h) If the forward response plot is linear and if the residual plot is rectangu-
lar about the horizontal axis, then the quadratic model may be appropriate.
Comment on the two plots.
Chapter 2
25
CHAPTER 2. THE LOCATION MODEL 26
60
1000
height
40
500
20
0
0
Index height
1500
1000
dens$y
0.002
500
0.0
dens$x
Figure 2.1: Dot plot, histogram, density estimate, and box plot for heights
from Buxton (1920).
CHAPTER 2. THE LOCATION MODEL 27
heights around 0.75 inches! It appears that their heights were recorded under
the variable “head length,” so these height outliers can be corrected. Note
that the presence of outliers is easily detected in all four plots.
Point estimation is one of the oldest problems in statistics and four of
the most important statistics for the location model are the sample mean,
median, variance, and the median absolute deviation (mad). Let Y1 , . . . , Yn
be the random sample; ie, assume that Y1 , ..., Yn are iid.
Definition 2.1. The sample mean
n
Yi
Y = i=1 . (2.2)
n
population sample
E(Y ), µ, θ Y n , E(n) µ̂, θ̂
MED(Y ), M MED(n), M̂
VAR(Y ), σ 2 VAR(n), S 2 , σ̂ 2
SD(Y ), σ SD(n), S, σ̂
MAD(Y ) MAD(n)
IQR(Y ) IQR(n)
They are also quite old. Rey (1978, p. 2) quotes Thucydides on a technique
used by Greek besiegers in the winter of 428 B.C. Cities were often surrounded
by walls made of layers of bricks, and besiegers made ladders to scale these
walls. The length of the ladders was determined by counting the layers of
bricks. Many soldiers counted the number of bricks, and the mode of the
counts was used to estimate the number of layers. The reasoning was that
some of the counters would make mistakes, but the majority were likely to
hit the true count. If the majority did hit the true count, then the sample
median would equal the mode. In a lecture, Professor Portnoy stated that in
215 A.D., an “eggs bulk” of impurity was allowed in the ritual preparation of
food, and two Rabbis desired to know what is an “average sized egg” given
a collection of eggs. One said use the middle sized egg while the other said
average the largest and smallest eggs of the collection. Hampel, Ronchetti,
Rousseeuw and Stahel (1986, p. 65) attribute MAD(n) to Gauss in 1816.
and
P (Y ∈ (MED(Y ) − MAD(Y ), MED(Y ) + MAD(Y )) ≤ 0.5.
MAD(Y ) and MED(Y ) are often simple to find for location, scale, and
location–scale families. Assume that the cdf F of Y has a probability density
function (pdf) or probability mass function (pmf) f. The following definitions
are taken from Casella and Berger (2002, p. 116-119) and Lehmann (1983,
p. 20).
Definition 2.7. Let fY (y) be the pdf of Y. Then the family of pdf’s
fW (w) = fY (w − µ) indexed by the location parameter µ, −∞ < µ < ∞, is
CHAPTER 2. THE LOCATION MODEL 31
the location family for the random variable W = µ + Y with standard pdf
fY (y).
Definition 2.8. Let fY (y) be the pdf of Y. Then the family of pdf’s
fW (w) = (1/σ)fY (w/σ) indexed by the scale parameter σ > 0, is the scale
family for the random variable W = σY with standard pdf fY (y).
Definition 2.9. Let fY (y) be the pdf of Y. Then the family of pdf’s
fW (w) = (1/σ)fY ((w − µ)/σ) indexed by the location and scale parameters
µ, −∞ < µ < ∞, and σ > 0, is the location–scale family for the random
variable W = µ + σY with standard pdf fY (y).
Table 2.2 gives the population mads and medians for some “brand name”
distributions. The distributions are location–scale families except for the
exponential and tp distributions. The notation tp denotes a t distribution
with p degrees of freedom while tp,α is the α percentile of the tp distribution,
ie P (tp ≤ tp,α) = α. Hence tp,0.5 = 0 is the population median. The second
column of Table 2.2 gives the section of Chapter 3 where the random variable
is described further. For example, the exponential (λ) random variable is
described in Section 3.7. Table 2.3 presents approximations for the binomial,
chi-square, and gamma distributions.
Finding MED(Y ) and MAD(Y ) for symmetric distributions and location–
scale families is made easier by the following lemma and Table 2.2. Let
F (yα) = P (Y ≤ yα) = α for 0 < α < 1 where the cdf F (y) = P (Y ≤ y). Let
D = MAD(Y ), M = MED(Y ) = y0.5 and U = y0.75.
Lemma 2.1. a) If W = a + bY, then MED(W ) = a + bMED(Y ) and
MAD(W ) = |b|MAD(Y ).
b) If Y has a pdf that is continuous and positive on its support and
symmetric about µ, then MED(Y ) = µ and MAD(Y ) = y0.75 − MED(Y ).
Find M = MED(Y ) by solving the equation F (M) = 0.5 for M, and find U
by solving F (U) = 0.75 for U. Then D = MAD(Y ) = U − M.
c) Suppose that W is from a location–scale family with standard pdf
fY (y) that is continuous and positive on its support. Then W = µ + σY
where σ > 0. First find M by solving FY (M) = 0.5. After finding M, find
D by solving FY (M + D) − FY (M − D) = 0.5. Then MED(W ) = µ + σM
and MAD(W ) = σD.
Proof sketch. a) Assume the probability density function of Y is con-
CHAPTER 2. THE LOCATION MODEL 32
Table 2.2: MED(Y ) and MAD(Y ) for some useful random variables.
Let D = MAD(Z) and let P (Z ≤ z) = Φ(z) be the cdf of Z. Now Φ(z) does
not have a closed form but is tabled extensively. Lemma 2.1b) implies that
D = z0.75 − 0 = z0.75 where P (Z ≤ z0.75) = 0.75. From a standard normal
table, 0.67 < D < 0.68 or D ≈ 0.674. A more accurate value can be found
with the following R/Splus command.
> qnorm(0.75)
[1] 0.6744898
Hence MAD(W ) ≈ 0.6745σ.
Example 2.4. If W is exponential (λ), then the cdf of W is FW (w) =
1 − exp(−w/λ) for w > 0 and FW (w) = 0 otherwise. Since exp(log(1/2)) =
exp(− log(2)) = 0.5, MED(W ) = log(2)λ. Since the exponential distribution
is a scale family with scale parameter λ, MAD(W ) = Dλ for some D > 0.
Hence
0.5 = FW (log(2)λ + Dλ) − FW (log(2)λ − Dλ),
or 0.5 =
> tem(0.48)
[1] 0.997291
> tem(0.49)
[1] 1.01969
> tem(0.484)
[1] 1.006238
> tem(0.483)
[1] 1.004
> tem(0.481)
[1] 0.9995264
> tem(0.482)
[1] 1.001763
> tem(0.4813)
[1] 1.000197
> tem(0.4811)
[1] 0.99975
> tem(0.4812)
[1] 0.9999736
Hence D ≈ 0.4812 and MAD(W ) ≈ 0.4812λ ≈ λ/2.0781. If X is a
two parameter exponential (θ, λ) random variable, then X = θ + W. Hence
MED(X) = θ + log(2)λ and MAD(X) ≈ λ/2.0781.
Example 2.5. This example shows how to approximate the population
median and mad under severe contamination when the “clean” observations
are from a symmetric location–scale family. Let Φ be the cdf of the standard
normal, and let Φ(zα ) = α. Note that zα = Φ−1 (α). Suppose Y ∼ (1−γ)FW +
γFC where W ∼ N(µ, σ 2 ) and C is a random variable far to the right of µ.
Show a)
MED(Y ) ≈ µ + σz[ 1 ]
2(1−γ)
≈ 2σz[ 1
] .
2(1−γ)
and
MED(Y ) − µ 1
Φ( )≈ .
σ 2(1 − γ)
b) Since the mass of C is far to the right of µ,
MED(Y ) − MAD(Y ) − µ
= (1 − γ)[1 − Φ( )].
σ
Writing z[α] for zα gives
MED(Y ) − MAD(Y ) − µ 1 − 2γ
≈z .
σ 2(1 − γ)
Thus
1 − 2γ
MAD(Y ) ≈ MED(Y ) − µ − σz .
2(1 − γ)
Since z[α] = −z[1 − α],
1 − 2γ 1
−z =z
2(1 − γ) 2(1 − γ)
and
1 1
MAD(Y ) ≈ µ + σz − µ + σz .
2(1 − γ) 2(1 − γ)
Y ≈ N(νλ, νλ2 )
2
√ large ν. If X is N(µ, σ ) then MAD(X) ≈ σ/1.483. Hence MAD(Y ) ≈
for
λ√ν/1.483. Assuming that ν is large, solve MED(n) = λν and MAD(n) =
λ ν/1.483 for ν and λ obtaining
2
MED(n) (1.483MAD(n))2
ν̂ ≈ and λ̂ ≈ .
1.483MAD(n) MED(n)
c) Suppose that Y1 , ..., Yn are iid from an extreme value distribution, then
the cdf of Y is
y−θ
F (y) = exp[− exp(−( ))].
σ
This family is an asymmetric location-scale family. Since 0.5 = F (MED(Y )),
MED(Y ) = θ − σ log(log(2)) ≈ θ + 0.36651σ. Let D = MAD(Y ) if θ = 0
and σ = 1. Then 0.5 = F [MED(Y ) + MAD(Y )] − F [MED(Y ) − MAD(Y )].
Solving 0.5 = exp[− exp(−(0.36651 + D))] − exp[− exp(−(0.36651 − D))] for
D numerically yields D = 0.767049. Hence MAD(Y ) = 0.767049σ.
d) Sometimes MED(n) and MAD(n) can also be used to estimate the
parameters of two parameter families that are not location–scale families.
Suppose that Y1 , ..., Yn are iid from a Weibull(φ, λ) distribution where λ, y,
and φ are all positive. The cdf of Y is F (y) = 1 − exp(−y φ /λ) for y >
0. Taking φ = 1 gives theexponential(λ) distribution while φ = 2 gives
the Rayleigh(µ = 0, σ = λ/2) distribution. Since F (MED(Y )) = 1/2,
MED(Y ) = (λ log(2))1/φ . These results suggest that if φ is known, then
(MED(n))φ
λ̂ = .
log(2)
Falk (1997) shows that under regularity conditions, the joint distribution
of the sample median and mad is asymptotically normal. See Section 2.9.
CHAPTER 2. THE LOCATION MODEL 37
Table 2.4: Robust point estimators for some useful random variables.
BIN(k,ρ) ρ̂ ≈ MED(n)/k
C(µ, σ) µ̂ = MED(n) σ̂ = MAD(n)
χ2p p̂ ≈ MED(n) + 2/3, rounded
DE(θ, λ) θ̂ = MED(n) λ̂ = 1.443MAD(n)
EXP(λ) λ̂1 = 1.443MED(n) λ̂2 = 2.0781MAD(n)
EXP(θ, λ) θ̂ = MED(n) − 1.440MAD(n) λ̂ = 2.0781MAD(n)
EV(θ, σ) θ̂ = MED(n) − 0.4778MAD(n) σ̂ = 1.3037MAD(n)
2
G(ν, λ) ν̂ ≈ [MED(n)/1.483MAD(n)]2 λ̂ ≈ [1.483MAD(n)]
MED(n)
HN(µ, σ) µ̂ = MED(n) − 1.6901MAD(n) σ̂ = 2.5057MAD(n)
L(µ, σ) µ̂ = MED(n) σ̂ = 0.9102MAD(n)
N(µ, σ 2) µ̂ = MED(n) σ̂ = 1.483MAD(n)
R(µ, σ) µ̂ = MED(n) − 2.6255MAD(n) σ̂ = 2.230MAD(n)
U(θ1, θ2 ) θˆ1 = MED(n) − 2MAD(n) ˆ
θ2 = MED(n) + 2MAD(n)
where
2 1
σM = ,
4[f(MED(Y ))]2
and
2 1 3 2 3 1
σD = 2
− + 2
= .
64 [f(ξ3/4)] f(ξ3/4 )f(ξ1/4) [f(ξ1/4 )] 16[f(ξ3/4 )]2
CHAPTER 2. THE LOCATION MODEL 38
Three common approximate level α tests of hypotheses all use the null
hypothesis Ho : µW = µo . A right tailed test uses the alternative hypothesis
HA : µW > µo , a left tailed test uses HA : µW < µo , and a two tail test uses
HA : µW = µo . The test statistic is
Wn − µo
to = ,
SE(Wn )
and the (approximate) p-values are P (Z > to) for a right tail test, P (Z < to )
for a left tail test, and 2P (Z > |to|) = 2P (Z < −|to|) for a two tail test. The
null hypothesis Ho is rejected if the p-value < α.
Remark 2.1. Frequently the large sample CIs and tests can be improved
for smaller samples by substituting a t distribution with p degrees of freedom
for the standard normal distribution Z where p ≡ pn is some increasing
function of the sample size n. Then the 100(1 − α)% CI for µW is given by
The test statistic rarely has an exact tp distribution, but the approximation
tends to make the CI’s and tests more conservative; ie, the CI’s are longer
and Ho is less likely to be rejected. This book will typically use very simple
rules for p and not investigate the exact distribution of the test statistic.
Paired and two sample procedures can be obtained directly from the one
sample procedures. Suppose there are two samples Y1 , ..., Yn and X1 , ..., Xm.
If n = m and it is known that (Yi , Xi ) match up in correlated pairs, then
paired CIs and tests apply the one sample procedures to the differences Di =
Yi − Xi . Otherwise, assume the two samples are independent, that n and m
are large, and that
√ 2
n(W n (Y ) − µW (Y )) 0 σ (Y ) 0
√ D
→N , W
2 .
m(Wm (X) − µW (X)) 0 0 σW (X)
Then
2
(Wn (Y ) − µW (Y )) 0 σW (Y )/n 0
≈N , 2 ,
(Wm (X) − µW (X)) 0 0 σW (X)/m
and
2 2
σW (Y ) σW (X)
Wn (Y ) − Wm (X) − (µW (Y ) − µW (X)) ≈ N(0, + ).
n m
Hence
2 2
SW (Y ) SW (X)
SE(Wn (Y ) − Wm (X)) = + ,
n m
and the large sample 100(1 − α)% CI for µW (Y ) − µW (X) is given by
(Wn (Y ) − Wm (X)) ± z1−α/2SE(Wn (Y ) − Wm (X)).
Often approximate level α tests of hypotheses use the null hypothesis
Ho : µW (Y ) = µW (X). A right tailed test uses the alternative hypothesis
HA : µW (Y ) > µW (X), a left tailed test uses HA : µW (Y ) < µW (X), and a
two tail test uses HA : µW (Y ) = µW (X). The test statistic is
Wn (Y ) − Wm (X)
to = ,
SE(Wn (Y ) − Wm (X))
and the (approximate) p-values are P (Z > to) for a right tail test, P (Z < to )
for a left tail test, and 2P (Z > |to|) = 2P (Z < −|to|) for a two tail test. The
null hypothesis Ho is rejected if the p-value < α.
CHAPTER 2. THE LOCATION MODEL 43
Y n − Xm
to = .
2 (Y )
Sn 2 (X )
Sm
n
+ m
The right tailed p-value is given by P (tp > to ). For sample means, values of
the degrees of freedom that are more accurate than p = min(n − 1, m − 1)
can be computed. See Moore (2004, p. 452).
CHAPTER 2. THE LOCATION MODEL 44
1
Un
Wn = Wn (Ln , Un ) = [Ln Y(Ln +1) + Y(i) + (n − Un )Y(Un ) ]. (2.13)
n
i=Ln +1
1
Un
Rn = Rn (Ln , Un ) = Y(i) (2.14)
Un − Ln
i=Ln +1
[θ̂n − k1 Dn , θ̂n + k2 Dn ]
points estimate lower and upper population percentiles L(F ) and U(F ) and
change with the distribution F.
Two stage estimators are frequently used in robust statistics. Often the
initial estimator used in the first stage has good resistance properties but
has a low asymptotic relative efficiency or no convenient formula for the SE.
Ideally, the estimator in the second stage will have resistance similar to the
initial estimator but will be efficient and easy to use. The metrically trimmed
mean Mn with tuning parameter k1 = k2 ≡ k = 6 will often be the initial
estimator for the two stage trimmed means. That is, retain the cases that
fall in the interval
Let L(Mn ) be the number of observations that fall to the left of MED(n) −
k1 MAD(n) and let n − U(Mn ) be the number of observations that fall to
the right of MED(n) + k2 MAD(n). When k1 = k2 ≡ k ≥ 1, at least half of
the cases will be covered. Consider the set of 51 trimming proportions in the
set C = {0, 0.01, 0.02, ..., 0.49, 0.50}. Alternatively, the coarser set of 6 trim-
ming proportions C = {0, 0.01, 0.1, 0.25, 0.40, 0.49} may be of interest. The
greatest integer function (eg 7.7
= 7) is used in the following definitions.
Definition 2.14. Consider the smallest proportion αo,n ∈ C such that
αo,n ≥ L(Mn )/n and the smallest proportion 1 − βo,n ∈ C such that 1 −
βo,n ≥ 1 − (U(Mn )/n). Let αM,n = max(αo,n , 1 − βo,n ). Then the two stage
symmetrically trimmed mean TS,n is the αM,n trimmed mean. Hence TS,n
is a randomly trimmed mean with Ln = n αM,n
and Un = n − Ln . If
αM,n = 0.50, then use TS,n = MED(n).
Definition 2.15. As in the previous definition, consider the smallest
proportion αo,n ∈ C such that αo,n ≥ L(Mn )/n and the smallest proportion
1 − βo,n ∈ C such that 1 − βo,n ≥ 1 − (U(Mn )/n). Then the two stage asym-
metrically trimmed mean TA,n is the (αo,n , 1 − βo,n ) trimmed mean. Hence
TA,n is a randomly trimmed mean with Ln = n αo,n
and Un = n βo,n
.
If αo,n = 1 − βo,n = 0.5, then use TA,n = MED(n).
Example 2.11. These two stage trimmed means are almost as easy to
compute as the classical trimmed mean, and no knowledge of the unknown
parameters is needed to do inference. First, order the data and find the
number of cases L(Mn ) less than MED(n) − k1 MAD(n) and the number of
cases n − U(Mn ) greater than MED(n) + k2 MAD(n). (These are the cases
CHAPTER 2. THE LOCATION MODEL 46
Remark 2.3. A simple method for computing VSW (Ln , Un ) has the
following steps. First, find d1 , ..., dn where
Y(Ln +1), i ≤ Ln
di = Y(i) , Ln + 1 ≤ i ≤ Un
Y(Un ) , i ≥ Un + 1.
Then the Winsorized variance is the sample variance Sn2 (d1 , ..., dn) of d1 , ..., dn,
and the scaled Winsorized variance
Sn2 (d1 , ..., dn)
VSW (Ln , Un ) = . (2.16)
([Un − Ln ]/n)2
Notice that the SE given in Definition 2.16 is the SE for the δ trimmed mean
where Ln and Un are fixed constants rather than random.
Application 2.4. Let Tn be the two stage (symmetrically or) asymmet-
rically trimmed mean that trims the Ln smallest cases and the n − Un largest
cases. Then for the one and two sample procedures described in Section 2.5,
use the one sample standard error SERM (Ln , Un ) given in Definition 2.16 and
the tp distribution where the degrees of freedom p = Un − Ln − 1.
CHAPTER 2. THE LOCATION MODEL 47
The CI’s and tests for the α trimmed mean and two stage trimmed means
given by Applications 2.3 and 2.4 are very similar once Ln has been computed.
For example, a large sample 100 (1 − γ)% confidence interval (CI) for µT is
(Tn − tUn −Ln −1,1− γ2 SERM (Ln , Un ), Tn + tUn−Ln −1,1− γ2 SERM (Ln , Un )) (2.17)
for every
> 0, and Xn converges almost everywhere (or almost surely, or
with probability 1) to X if
P ( lim Xn = X) = 1.
n→∞
for a ≤ y ≤ b. Also G is 0 for y < a and G is 1 for y > b. The mean and
variance of YT are
∞ b
ydF (y)
µT = µT (a, b) = ydG(y) = a (2.19)
−∞ β −α
and b
∞
y 2dF (y)
σT2 = σT2 (a, b) = (y − µT )2dG(y) = a
− µ2T .
−∞ β−α
See Cramér (1946, p. 247).
Definition 2.18. The Winsorized random variable
a, Y ≤ a
YW = YW (a, b) = Y, Y ≤ b
b, Y ≥ b.
and b
2 2 2 2
σW = σW (a, b) = αa + (1 − β)b + y 2dF (y) − µ2W .
a
P (Y < a) and let αo ∈ C = {0, 0.01, 0.02, ..., 0.49, 0.50} be the smallest value
in C such that αo ≥ α. Similarly, let β = F (b) and let 1−βo ∈ C be the small-
est value in the index set C such that 1 − βo ≥ 1 − β. Let αo = F (ao−), and
let βo = F (bo). Recall that L(Mn ) is the number of cases trimmed to the left
and that n − U(Mn ) is the number of cases trimmed to the right by the met-
rically trimmed mean Mn . Let αo,n ≡ α̂o be the smallest value in C such that
αo,n ≥ L(Mn )/n, and let 1−βo,n ≡ 1− β̂o be the smallest value in C such that
1−βo,n ≥ 1−(U(Mn )/n). Then the robust estimator TA,n is the (αo,n , 1−βo,n )
trimmed mean while TS,n is the max(αo,n , 1 − βo,n )100% trimmed mean. The
following lemma is useful for showing that TA,n is asymptotically equivalent
to the (αo , 1 − βo ) trimmed mean and that TS,n is asymptotically equivalent
to the max(αo , 1 − βo) trimmed mean.
Lemma 2.4: Shorack and Wellner (1986, p. 682-683). Let F
have a strictly positive and continuous derivative in some neighborhood of
MED(Y ) ± kMAD(Y ). Assume that
√
n(MED(n) − MED(Y )) = OP (1) (2.23)
and √
n(MAD(n) − MAD(X)) = OP (1). (2.24)
Then
√ L(Mn )
n( − α) = OP (1) (2.25)
n
and
√ U(Mn )
n( − β) = OP (1). (2.26)
n
Corollary 2.5. Let Y1 , ..., Yn be iid from a distribution with cdf F that
has a strictly positive and continuous pdf f on its support. Let αM =
max(αo , 1 − βo ) ≤ 0.49, βM = 1 − αM , aM = F −1(αM ), and bM = F −1(βM ).
Assume that α and 1 − β are not elements of C = {0, 0.01, 0.02, ..., 0.50}.
Then
√ D σ 2 (ao , bo )
n[TA,n − µT (ao , bo )] → N(0, W ),
(βo − αo )2
and
√ D σ 2 (aM , bM )
n[TS,n − µT (aM , bM )] → N(0, W ).
(βM − αM )2
CHAPTER 2. THE LOCATION MODEL 53
Proof. The first result follows from Theorem 2.2 if the probability that
TA,n is the (αo , 1 − βo) trimmed mean goes to one as n tends to infinity. This
D D
condition holds if L(Mn )/n → α and U(Mn )/n → β. But these conditions
follow from Lemma 2.4. The proof for TS,n is similar. QED
Now
Y −T
ψk ( )=1
S
if T − kS ≤ Y ≤ T + kS and is zero otherwise (technically the derivative is
undefined at y = ± k, but assume that Y is a continuous random variable
so that the probability of a value occuring on a “corner” of the ψ function is
zero). Let Ln count the number of observations Yi < MED(n) − kMAD(n),
and let n − Un count the number of observations Yi > MED(n) + kMAD(n).
Set T (0) = MED(n) and S = MAD(n). Then
n
Yi − T (0)
ψk ( ) = Un − Ln .
i=1
S
Since
Yi − MED(n)
ψk ( )=
MAD(n)
−k, Yi < MED(n) − kMAD(n)
Ỹ , MED(n) − kMAD(n) ≤ Yi ≤ MED(n) + kMAD(n)
i
k, Yi > MED(n) + kMAD(n),
where Ỹi = (Yi − MED(n))/MAD(n),
n
Y(i) − T (0)
Un
Y(i) − T (0)
ψk ( ) = −kLn + k(n − Un ) + .
S S
i=1 i=Ln +1
Hence n −MED(n)
i=1 ψk ( YiMAD(n)
)
MED(n) + S
ψk ( Yi −MED(n) )
n
MAD(n)
i=1
n
kMAD(n)(n − Un − Ln ) + Ui=L n +1
[Y(i) − MED(n)]
= MED(n) + ,
Un − Ln
and Huber’s one step M-estimator
n
kMAD(n)(n − Un − Ln ) + Ui=L n +1
Y(i)
H1,n = .
Un − Ln
CHAPTER 2. THE LOCATION MODEL 56
and
Similarly
and thus
or
|MAD(n) − MD(n)| ≤ |MED(n) − MED(Y )|. (2.27)
Adding and subtracting MAD(Y ) to the left hand side shows that
The main point of the following theorem is that the joint distribution of
MED(n) and MAD(n) is asymptotically normal. Hence the limiting distribu-
tion of MED(n) + kMAD(n) is also asymptotically normal for any constant
k. The parameters of the covariance matrix are quite complex and hard to es-
timate. The assumptions of f used in Theorem 2.8 guarantee that MED(Y )
and MAD(Y ) are unique.
Theorem 2.8: Falk (1997). Let the cdf F of Y be continuous near and
differentiable at MED(Y ) = F −1(1/2) and MED(Y )±MAD(Y ). Assume that
f = F , f(F −1 (1/2)) > 0, and A ≡ f(F −1(1/2) − MAD(Y )) + f(F −1 (1/2) +
MAD(Y )) > 0. Let C ≡ f(F −1 (1/2) − MAD(Y )) − f(F −1(1/2) + MAD(Y )),
and let B ≡ C 2 +4Cf(F −1(1/2))[1−F (F −1 (1/2)−MAD(Y ))−F (F −1(1/2)+
MAD(Y ))]. Then
√ MED(n) MED(Y ) D
n − →
MAD(n) MAD(Y )
2
0 σM σM,D
N , 2 (2.29)
0 σM,D σD
where
2 1 2 1 B
σM = , σD = (1 + 2 −1 1 ),
4f 2 (F −1( 12 )) 4A 2 f (F ( 2 ))
and
1 −1 1 C
σM,D = 1 (1 − 4F (F ( ) + MAD(Y )) + ).
−1
4Af(F ( 2 )) 2 f(F ( 12 ))
−1
Determining whether the population median and mad are unique can be
useful. Recall that F (y) = P (Y ≤ y) and F (y−) = P (Y < y). The median
is unique unless there is a flat spot at F −1(0.5), that is, unless there exist a
and b with a < b such that F (a) = F (b) = 0.5. MAD(Y ) may be unique even
if MED(Y ) is not, see Problem 2.7. If MED(Y ) is unique, then MAD(Y )
is unique unless F has flat spots at both F −1 (MED(Y ) − MAD(Y )) and
F −1(MED(Y ) + MAD(Y )). Moreover, MAD(Y ) is unique unless there exist
a1 < a2 and b1 < b2 such that F (a1) = F (a2), F (b1) = F (b2),
P (ai ≤ Y ≤ bi) = F (bi) − F (ai−) ≥ 0.5,
and
P (Y ≤ ai ) + P (Y ≥ bi ) = F (ai) + 1 − F (bi−) ≥ 0.5
CHAPTER 2. THE LOCATION MODEL 58
for i = 1, 2. The following lemma gives some simple bounds for MAD(Y ).
Lemma 2.9. Assume MED(Y ) and MAD(Y ) are unique. a) Then
min{MED(Y ) − F −1(0.25), F −1 (0.75) − MED(Y )} ≤ MAD(Y ) ≤
max{MED(Y ) − F −1(0.25), F −1 (0.75) − MED(Y )}. (2.30)
b) If Y is symmetric about µ = F −1(0.5), then the three terms in a) are
equal.
c) If the distribution is symmetric about zero, then MAD(Y ) = F −1 (0.75).
d) If Y is symmetric and continuous with a finite second moment, then
MAD(Y ) ≤ 2VAR(Y ).
e) Suppose Y ∈ [a, b]. Then
0 ≤ MAD(Y ) ≤ m = min{MED(Y ) − a, b − MED(Y )} ≤ (b − a)/2,
and the inequalities are sharp.
Proof. a) This result follows since half the mass is between the upper
and lower quartiles and the median is between the two quartiles.
b) and c) are corollaries of a).
d) This inequality holds by Chebyshev’s inequality, since
P ( |Y − E(Y )| ≥ MAD(Y ) ) = 0.5 ≥ P ( |Y − E(Y )| ≥ 2VAR(Y ) ),
and E(Y ) = MED(Y ) for symmetric distributions with finite second mo-
ments.
e) Note that if MAD(Y ) > m, then either MED(Y ) − MAD(Y ) < a
or MED(Y ) + MAD(Y ) > b. Since at least half of the mass is between a
and MED(Y ) and between MED(Y ) and b, this contradicts the definition of
MAD(Y ). To see that the inequalities are sharp, note that if at least half of
the mass is at some point c ∈ [a, b], than MED(Y ) = c and MAD(Y ) = 0.
If each of the points a, b, and c has 1/3 of the mass where a < c < b, then
MED(Y ) = c and MAD(Y ) = m. QED
Many other results for MAD(Y ) and MAD(n) are possible. For example,
note that Lemma 2.9 b) implies that when Y is symmetric, MAD(Y ) =
F −1(3/4) − µ and F (µ + MAD(Y )) = 3/4. Also note that MAD(Y ) and the
interquartile range IQR(Y ) are related by
2MAD(Y ) = IQR(Y ) ≡ F −1 (0.75) − F −1 (0.25)
CHAPTER 2. THE LOCATION MODEL 59
2.10 Summary
1) Given a small data set, recall that
n
Yi
Y = i=1
n
and the sample variance
n n
2 − Y )2
i=1 (Yi Yi2 − n(Y )2
VAR(n) = S = Sn2 = = i=1
,
n−1 n−1
and the sample standard deviation (SD)
S = Sn = Sn2 .
CHAPTER 2. THE LOCATION MODEL 60
Y(n/2) + Y((n/2)+1)
MED(n) = if n is even.
2
The notation MED(n) = MED(Y1 , ..., Yn) will also be used. To find the
sample median, sort the data from smallest to largest and find the middle
value or values.
The sample median absolute deviation
To find MAD(n), find Di = |Yi − MED(n)|, then find the sample median
of the Di by ordering them from smallest to largest and finding the middle
value or values.
2) Find the population median M = MED(Y ) by solving the equation
F (M) = 0.5 for M where the cdf F (y) = P (Y ≤ y). If Y has a pdf f(y)
that is symmetric about µ, then M = µ. If W = a + bY, then MED(W ) =
a + bMED(Y ). Often a = µ and b = σ.
3) To find the population median absolute deviation D = MAD(Y ), first
find M = MED(Y ) as in 2) above.
a) Then solve F (M + D) − F (M − D) = 0.5 for D.
b) If Y has a pdf that is symmetric about µ, then let U = y0.75 where
P (Y ≤ yα) = α, and yα is the 100αth percentile of Y for 0 < α < 1.
Hence M = y0.5 is the 50th percentile and U is the 75th percentile. Solve
F (U) = 0.75 for U. Then D = U − M.
c) If W = a + bY, then MAD(W ) = |b|MAD(Y ).
MED(Y ) and MAD(Y ) need not be unique, but for “brand name” con-
tinuous random variables, they are unique.
4) A large sample 100 (1 − α)% confidence interval (CI) for θ is
θ̂ ± tp,1− α2 SE(θ̂)
CHAPTER 2. THE LOCATION MODEL 61
1
Un
Tn = Tn (Ln , Un ) = Y(i)
Un − Ln i=L +1
n
where Ln = n/4
and Un = n − Ln . That is, order the data, delete the
Ln smallest cases and the Ln largest cases and take the sample mean of
the remaining Un − Ln cases. The 25% trimmed mean is estimating the
population truncated mean
y0.75
µT = 2yfY (y)dy.
y0.25
2.11 Complements
Chambers, Cleveland, Kleiner and Tukey (1983) is an excellent source for
graphical procedures such as quantile plots, QQ-plots, and box plots.
The confidence intervals and tests for the sample median and 25% trimmed
mean can be modified for censored data as can the robust point estimators
based on MED(n) and MAD(n). Suppose that Y(R+1) , ..., Y(n) have been right
censored (similar results hold for left censored data). Then create a pseudo
sample Z(i) = Y(R) for i > R and Z(i) = Y(i) for i ≤ R. Then compute the
robust estimators based Z1 , ..., Zn . These estimators will be identical to the
estimators based on Y1 , ..., Yn (no censoring) if the amount of right censoring
is moderate. For a one parameter family, nearly half of the data can be right
censored if the estimator is based on the median. If the sample median and
MAD are used for a two parameter family, the proportion of right censored
data depends on the skewness of the distribution. Symmetric data can tol-
erate nearly 25% right censoring, right skewed data a larger percentage, and
left skewed data a smaller percentage. See Olive (2006). He and Fung (1999)
present an alternative robust method that also works well for censored data.
Huber (1981, p. 74-75) and Chen (1998) show that the sample median
minimizes the asymptotic bias for estimating MED(Y ) for the family of sym-
metric contaminated distributions, and Huber (1981) concludes that since
the asymptotic variance is going to zero for reasonable estimators, MED(n)
is the estimator of choice for large n. Hampel, Ronchetti, Rousseeuw, and
Stahel (1986, p. 133-134, 142-143) contains some other optimality properties
of MED(n) and MAD(n). Price and Bonnett (2001), McKean and Schrader
(1984) and Bloch and Gastwirth (1968) are useful references for estimating
the SE of the sample median.
Several other approximations for the standard error of the sample median
SE(MED(n)) could be used.
a) McKean and Schrader (1984) proposed
Y(n−c+1) − Y(c)
SE(MED(n)) =
2z1− α2
CHAPTER 2. THE LOCATION MODEL 63
where c = (n+1)/2 − z1−α/2 n/4 is rounded up to the nearest integer. This
− α)% CI
estimator was based on the half length of a distribution free 100 (1 √
(Y(c) , Y(n−c+1) ) for MED(Y ). Use the tp approximation with p = 2 n
− 1.
b) This proposal is also due to Bloch and Gastwirth (1968). Let Un =
n − Ln where Ln = n/2
− 0.5n0.8 and use
2.12 Problems
PROBLEMS WITH AN ASTERISK * ARE ESPECIALLY USE-
FUL.
2.1. Write the location model in matrix form.
2.2. Let fY (y) be the pdf of Y. If W = µ + Y where −∞ < µ < ∞, show
that the pdf of W is fW (w) = fY (w − µ).
2.3. Let fY (y) be the pdf of Y. If W = σY where σ > 0, show that the
pdf of W is fW (w) = (1/σ)fY (w/σ).
2.4. Let fY (y) be the pdf of Y. If W = µ + σY where −∞ < µ < ∞ and
σ > 0, show that the pdf of W is fW (w) = (1/σ)fY ((w − µ)/σ).
√
2.5. Use Theorem 2.8 to find the limiting distribution of n(MED(n) −
MED(Y )).
2.6. The interquartile range IQR(n) = ξˆn,0.75 − ξˆn,0.25 and is a popular
estimator of scale. Use Theorem 2.6 to show that
√ 1
n (IQR(n) − IQR(Y )) → N(0, σA2 )
D
2
where
1 3 2 3
σA2 = − + .
64 [f(ξ3/4 )]2 f(ξ3/4)f(ξ1/4 ) [f(ξ1/4)]2
2.7. Let the pdf of Y be f(y) = 1 if 0 < y < 0.5 or if 1 < y < 1.5. Assume
that f(y) = 0, otherwise. Then Y is a mixture of two uniforms, one U(0, 0.5)
and the other U(1, 1.5). Show that the population median MED(Y ) is not
unique but the population mad MAD(Y ) is unique.
√
2.8. a) Let Ln = 0 and Un = n. Prove that SERM (0, n) = S/ n. In other
words, the SE given by Definition 2.16 reduces to the SE for the sample mean
if there is no trimming.
b) Prove Remark 2.3:
Sn2 (d1 , ..., dn)
VSW (Ln , Un ) = .
[(Un − Ln )/n]2
CHAPTER 2. THE LOCATION MODEL 66
2.9. Find a 95% CI for µT based on the 25% trimmed mean for the
following data sets. Follow Examples 2.12 and 2.13 closely with Ln = 0.25n
and Un = n − Ln .
a) 6, 9, 9, 7, 8, 9, 9, 7
b) 66, 99, 9, 7, 8, 9, 9, 7
2.10. Consider the data set 6, 3, 8, 5, and 2. Show work.
a) Find the sample mean Y .
b) Find the standard deviation S
c) Find the sample median MED(n).
d) Find the sample median absolute deviation MAD(n).
2.11∗. The Cushny and Peebles data set (see Staudte and Sheather 1990,
p. 97) is listed below.
1.2 2.4 1.3 1.3 0.0 1.0 1.8 0.8 4.6 1.4
2.12∗. Consider the following data set on Spring 2004 Math 580 home-
work scores.
66.7 76.0 89.7 90.0 94.0 94.0 95.0 95.3 97.0 97.7
2.13∗. Consider the following data set on Spring 2004 Math 580 home-
work scores.
66.7 76.0 89.7 90.0 94.0 94.0 95.0 95.3 97.0 97.7
to simulate data similar to the Buxton heights. Make a plot similar to Figure
2.1 using the following R/Splus commands.
> par(mfrow=c(2,2))
> plot(height)
> title("a) Dot plot of heights")
> hist(height)
> title("b) Histogram of heights")
> length(height)
[1] 87
CHAPTER 2. THE LOCATION MODEL 68
b) Compute the mean and 25% trimmed mean of 10000 simulated EXP(1)
random variables with the following commands.
y <- rexp(10000)
mean(y)
tmn(y)
metmn <-
function(x, k = 6)
{
madd <- mad(x, constant = 1)
med <- median(x)
mean(x[(x >= med - k * madd) & (x <= med + k * madd)])
}
y <- rnorm(10000)
metmn(y)
y <- rnorm(10000)
ratmn(y)
2.20. Download the R/Splus function rstmn that computes the two stage
symmetrically trimmed mean TS,n . Compute the TS,n for 10000 simulated
N(0, 1) random variables with the following commands.
CHAPTER 2. THE LOCATION MODEL 70
y <- rnorm(10000)
rstmn(y)
2.21∗. a) Download the cci function which produces a classical CI. The
default is a 95% CI.
b) Compute a 95% CI for the artificial height data set created in Problem
2.15. Use the command cci(height).
2.22∗. a) Download the R/Splus function medci that produces a CI using
the median and the Bloch and Gastwirth SE.
b) Compute a 95% CI for the artificial height data set created in Problem
2.15. Use the command medci(height).
2.23∗. a) Download the R/Splus function tmci that produces a CI using
the 25% trimmed mean as a default.
b) Compute a 95% CI for the artificial height data set created in Problem
2.15. Use the command tmci(height).
2.24. a) Download the R/Splus function atmci that produces a CI using
TA,n .
b) Compute a 95% CI for the artificial height data set created in Problem
2.15. Use the command atmci(height).
2.25. a) Download the R/Splus function stmci that produces a CI using
TS,n .
b) Compute a 95% CI for the artificial height data set created in Problem
2.15. Use the command stmci(height).
2.26. a) Download the R/Splus function stmci that produces a CI using
the median and SERM (Ln , Un ).
b) Compute a 95% CI for the artificial height data set created in Problem
2.15. Use the command med2ci(height).
2.27. a) Download the R/Splus function cgci that produces a CI using
TS,n and the coarse grid C = {0, 0.01, 0.1, 0.25, 0.40, 0.49}.
b) Compute a 95% CI for the artificial height data set created in Problem
2.15. Use the command cgci(height).
2.28. a) Bloch and Gastwirth (1968) suggest using
√
n
SE(MED(n)) = [Y([n/2]+m) − Y([n/2]−m) ]
4m
CHAPTER 2. THE LOCATION MODEL 71
qplot<-
function(y)
{ plot(sort(y), ppoints(y))
title("QPLOT")}
b) Make a Q plot of the height data from Problem 2.15 with the following
command.
qplot(height)
Y <- rnorm(1000)
qplot(y)
Chapter 3
72
CHAPTER 3. SOME USEFUL DISTRIBUTIONS 73
MED(Y ) ≈ p−2/3. See Pratt (1968, p. 1470) for more terms in the expansion
of MED(Y ).
Empirically, √
2p 2 √
MAD(Y ) ≈ (1 − )2 ≈ 0.9536 p.
1.483 9p
Note that p ≈ MED(Y ) + 2/3, and VAR(Y ) ≈ 2MED(Y ) + 4/3. Let i be an
integer such that i ≤ w < i + 1. Then define rnd(w) = i if i ≤ w ≤ i + 0.5
and rnd(w) = i + 1 if i + 0.5 < w < i + 1. Then p ≈ rnd(MED(Y ) + 2/3),
and the approximation can be replaced by equality for p = 1, . . . , 100.
There are several normal approximations for this distribution. For p large,
Y ≈ N(p, 2p), and √
2Y ≈ N( 2p, 1).
Let
α = P (Y ≤ χ2p,α) = Φ(zα)
where Φ is the standard normal cdf. Then
1
χ2p,α ≈ (zα + 2p)2.
2
The Wilson–Hilferty approximation is
13
Y 2 2
≈ N(1 − , ).
p 9p 9p
where χ2p,α and zα are the α percentiles of the χ2p and standard normal dis-
tributions, respectively. See Patel, Kapadia and Owen (1976, p. 194).
A trimming rule is keep yi if
2.0
yi ∈ [med(n) ± 10.0(1 + )mad(n)].
n
Note that F (θ + λ log(1000)) = 0.9995 ≈ F (MED(Y ) + 10.0MAD(Y )).
where P (Y ≤ χ22n, α ) = α/2 if Y is χ22n. See Patel, Kapadia and Owen (1976,
2
p. 188).
If all the yi ≥ 0, then the trimming rule is keep yi if
c2
0.0 ≤ yi ≤ 9.0(1 + )med(n)
n
where c2 = 2.0 seems to work well. Note that P (Y ≤ 9.0MED(Y )) ≈ 0.998.
MED(Y ) = θ + λ log(2)
and
MAD(Y ) ≈ λ/2.0781.
Hence θ ≈ MED(Y ) − 2.0781 log(2)MAD(Y ). See Rousseeuw and Croux
(1993) for similar results. Note that 2.0781 log(2) ≈ 1.44.
A trimming rule is keep yi if
c4
med(n) − 1.44(1.0 + )mad(n) ≤ yi ≤
n
c2
med(n) − 1.44mad(n) + 9.0(1 + )med(n)
n
CHAPTER 3. SOME USEFUL DISTRIBUTIONS 80
= 0.5[−e−MAD/λ + eMAD/λ]
assuming λ log(2) > MAD. Plug in MAD = λ/2.0781 to get the result.
1 y−θ y−θ
f(y) = exp(−( )) exp[− exp(−( ))]
σ σ σ
where y and θ are real and σ > 0. (Then −Y has an extreme value distribution
for the min or the log–Weibull distribution, see Problem 3.10.)
The cdf of Y is
y−θ
F (y) = exp[− exp(−( ))].
σ
This family is an asymmetric location–scale family with a mode at θ.
The mgf m(t) = exp(tθ)Γ(1 − σt) for |t| < 1/σ.
E(Y ) ≈ θ + 0.57721σ, and
VAR(Y ) = σ 2π 2 /6 ≈ 1.64493σ 2 .
and
MAD(Y ) ≈ 0.767049σ.
W = exp(−(Y − θ)/σ) ∼ EXP(1).
A trimming rule is keep yi if
y ν−1 e−y/λ
f(y) =
λν Γ(ν)
where ν, λ, and y are positive.
The mgf of Y is ν ν
1/λ 1
m(t) = 1 =
λ
−t 1 − λt
for t < 1/λ. The chf ν
1
c(t) = .
1 − iλt
E(Y ) = νλ.
VAR(Y ) = νλ2 .
λr Γ(r + ν)
E(Y r ) = if r > −ν.
Γ(ν)
Chen and Rubin (1986) show that λ(ν − 1/3) < MED(Y ) < λν = E(Y ).
Empirically, for ν > 3/2,
and √
λ ν
MAD(Y ) ≈ .
1.483
This family is a scale family for fixed ν, so if Y is G(ν, λ) then cY is G(ν, cλ)
for c > 0. If W is EXP(λ) then W is G(1, λ). If W is χ2p , then W is G(p/2, 2).
If Y and W are independent and Y is G(ν, λ) and W is G(φ, λ), then Y + W
is G(ν + φ, λ). Some classical estimates are given next. Let
yn
w = log
geometric mean(n)
Also
0.5000876 + 0.1648852w − 0.0544274w2
ν̂M LE ≈
w
for 0 < w ≤ 0.5772, and
for 0.5772 < w ≤ 17. If w > 17 then estimation is much more difficult, but a
rough approximation is ν̂ ≈ 1/w for w > 17. See Bowman and Shenton (1988,
p. 46) and Greenwood and Durand (1960). Finally, λ̂ = yn /ν̂. Notice that λ̂
may not be very good if ν̂ < 1/17. For some M–estimators, see Marazzi and
Ruffieux (1996).
Several normal approximations are available. For large ν, Y ≈ N(νλ, νλ2 ).
The Wilson–Hilferty approximation says that for ν ≥ 0.5,
1/3 1/3 1 2/3 1
Y ≈ N (νλ) (1 − ), (νλ) .
9ν 9ν
α = P [Y ≤ Gα ],
then
3
1 1
Gα ≈ νλ zα +1−
9ν 9ν
where zα is the standard normal percentile, α = Φ(zα ). Bowman and Shenton
(1988, p. 101) include higher order terms.
Next we give some trimming rules. Assume each yi > 0. Assume ν ≥ 0.5.
Rule 1. Assume β is known. Let ν̂ = (med(n)/λ) + (1/3). Keep yi if
yi ∈ [lo, hi] where
1 1
lo = max(0, ν̂λ [−(3.5 + 2/n) + 1 − ]3),
9ν̂ 9ν̂
and
1 1
hi = ν̂λ [(3.5 + 2/n) + 1 − ]3 .
9ν̂ 9ν̂
CHAPTER 3. SOME USEFUL DISTRIBUTIONS 83
where
c ∈ [9, 15].
2 −(y − µ)2
f(y) = √ exp ( )
2π σ 2σ 2
where σ > 0 and y ≥ µ and µ is real. Let Φ(y) denote the standard normal
cdf. Then the cdf of Y is
y−µ
F (y) = 2Φ( )−1
σ
for y > µ and F (y) = 0, otherwise. This is an asymmetric location–scale
same distribution as µ + σ|Z| where Z ∼ N(0, 1).
family that has the
E(Y ) = µ + σ 2/π ≈ µ + 0.797885σ.
2
VAR(Y ) = σ (π−2)
π
≈ 0.363380σ 2 .
Note that Z ∼ χ21. Hence the formula for the rth moment of the χ21
2
and
Φ(y) ≈ 0.5(1 + 1 − exp(−2y 2 /π) ).
CHAPTER 3. SOME USEFUL DISTRIBUTIONS 86
1 1
n n
Y n = µ̂ = Yi and S 2 = SY2 = σ̂ 2 = (Yi − Y n )2.
n i=1 n − 1 i=1
SY SY
(Y n − tn−1,1− α2 √ , Y n + tn−1,1− α2 √ )
n n
co + c1 m + c2 m 2
zα ≈ m −
1 + d1 m + d2 m2 + d3 m3
where
m = [−2 log(1 − α)]1/2,
c0 = 2.515517, c1 = 0.802853, c2 = 0.010328, d1 = 1.432788, d2 = 0.189269,
d3 = 0.001308, and 0.5 ≤ α. For 0 < α < 0.5,
zα = −z1−α .
CHAPTER 3. SOME USEFUL DISTRIBUTIONS 87
r σr
E(Y ) = for r < 1/λ.
1 − λr
MED(Y ) = σ2λ.
X = log(Y/σ) is EXP(λ) and W = log(Y ) is EXP(θ = log(σ), λ). Let
θ̂ = MED(W1, ..., Wn) − 1.440MAD(W1, ..., Wn). Then robust estimators are
σ̂ = eθ̂ and λ̂ = 2.0781MAD(W1 , ..., Wn).
A trimming rule is keep yi if
med(n) − 1.44mad(n) ≤ wi ≤ 10med(n) − 1.44mad(n)
where med(n) and mad(n) are applied to w1 , . . . , wn with wi = log(yi).
CHAPTER 3. SOME USEFUL DISTRIBUTIONS 88
e−θ θy
P (Y = y) =
y!
for y = 0, 1, . . . , where θ > 0.
The mgf of Y is m(t) = exp(θ(et − 1)), and the chf of Y is
c(t) = exp(θ(eit − 1)).
E(Y ) = θ, and Chen and Rubin (1986) and Adell and Jodrá (2005) show
that
−1 < MED(Y ) − E(Y ) < 1/3.
VAR(Y ) = θ.
The classical estimator of θ is θ̂ = Y n . √ √
The approximations Y ≈ N(θ, θ) and 2 Y ≈ N(2 θ, 1) are sometimes used.
Suppose each yi is a nonnegative integer. Then a trimming rule is keep yi if
√
wi = 2 yi is kept when a normal trimming rule is applied to the wi s. (This
rule can be very bad if the normal approximation is not good.)
where σ > 0, µ is real, and y ≥ µ. See Cohen and Whitten (1988, Ch. 10).
This is an asymmetric location–scale family.
The cdf of Y is 2
1 y−µ
F (y) = 1 − exp −
2 σ
for y ≥ µ, andF (y) = 0, otherwise.
E(Y ) = µ + σ π/2 ≈ µ + 1.253314σ.
VAR(Y ) = σ 2(4 − π)/2 ≈ 0.429204σ 2 .
MED(Y ) = µ + σ log(4) ≈ µ + 1.17741σ.
Hence µ ≈ MED(Y ) − 2.6255MAD(Y ) and σ ≈ 2.230MAD(Y ).
Let σD = MAD(Y ). If µ = 0, and σ = 1, then
0.5 = exp[−0.5( log(4) − D)2 ] − exp[−0.5( log(4) + D)2 ].
Γ( p+1
2
) y 2 −( p+1 )
f(y) = (1 + ) 2
(pπ)1/2Γ(p/2) p
where p is a positive integer and y is real. This family is symmetric about
0. The t1 distribution is the Cauchy(0, 1) distribution. If Z is N(0, 1) and is
independent of W ∼ χ2p, then
Z
W 1/2
(p)
CHAPTER 3. SOME USEFUL DISTRIBUTIONS 90
is tp.
E(Y ) = 0 for p ≥ 2.
MED(Y ) = 0.
VAR(Y ) = p/(p − 2) for p ≥ 3, and
MAD(Y ) = tp,0.75 where P (tp ≤ tp,0.75) = 0.75.
If α = P (tp ≤ tp,α), then Cooke, Craven, and Clarke (1982, p. 84) suggest
the approximation
w2
tp,α ≈ p[exp( α ) − 1)]
p
where
zα(8p + 3)
wα = ,
8p + 1
zα is the standard normal cutoff: α = Φ(zα ), and 0.5 ≤ α. If 0 < α < 0.5,
then
tp,α = −tp,1−α.
This approximation seems to get better as the degrees of freedom increase.
A trimming rule for p ≥ 3 is keep yi if yi ∈ [±5.2(1 + 10/n)mad(n)].
where med(n) is applied to w1, . . . , wn with wi = eyi − 1. See Problem 3.8 for
robust estimators.
where λ, y, and φ are all positive. For fixed φ, this is a scale family in
σ = λ1/φ .
The cdf of Y is F (y) = 1 − exp(−y φ /λ) for y > 0.
E(Y ) = λ1/φ Γ(1 + 1/φ).
VAR(Y ) = λ2/φ Γ(1 + 2/φ) − (E(Y ))2 .
r
E(Y r ) = λr/φ Γ(1 + ) for r > −φ.
φ
MED(Y ) = (λ log(2))1/φ .
Note that
(MED(Y ))φ
λ= .
log(2)
Since W = Y φ is EXP(λ), if all the yi > 0 and if φ is known, then a cleaning
rule is keep yi if
2
0.0 ≤ wi ≤ 9.0(1 + )med(n)
n
where med(n) is applied to w1 , . . . , wn with wi = yiφ . See Olive (2006) and
Problem 3.10c for robust estimators of φ and λ.
3.23 Complements
Many of the distribution results used in this chapter came from Johnson and
Kotz (1970a,b) and Patel, Kapadia and Owen (1976). Cohen and Whitten
(1988), Ferguson (1967), Castillo (1988), Cramér (1946), Kennedy and Gentle
(1980), Lehmann (1983), Meeker and Escobar (1998), Bickel and Doksum
(1977), DeGroot (1975), Hastings and Peacock (1975) and Leemis (1986) also
have useful results on distributions. Also see articles in Kotz and Johnson
(1982ab, 1983ab, 1985ab, 1986, 1988ab) and Armitrage and Colton (1998a-
f). Often an entire book is devoted to a single distribution, see for example,
Bowman and Shenton (1988).
Many of the robust point estimators in this chapter are due to Olive
(2006). These robust estimators are usually inefficient, but can be used as
starting values for iterative procedures such as maximum likelihood and as a
quick check for outliers. These estimators can also be used to create a robust
fully efficient cross checking estimator. See He and Fung (1999).
If no outliers are present and the sample size is large, then the robust
and classical methods should give similar estimates. If the estimates differ,
CHAPTER 3. SOME USEFUL DISTRIBUTIONS 93
3.24 Problems
PROBLEMS WITH AN ASTERISK * ARE ESPECIALLY USE-
FUL.
3.1. Verify the formula for the cdf F for the following distributions.
a) Cauchy (µ, σ).
b) Double exponential (θ, λ).
c) Exponential (λ).
d) Logistic (µ, σ).
e) Pareto (σ, λ).
f) Power (λ).
g) Uniform (θ1, θ2 ).
h) Weibull W (φ, λ).
3.2∗. Verify the formula for MED(Y ) for the following distributions.
a) Exponential (λ).
b) Lognormal (µ, σ 2). (Hint: Φ(0) = 0.5.)
c) Pareto (σ, λ).
d) Power (λ).
e) Uniform (θ1 , θ2).
f) Weibull (φ, λ).
3.3∗. Verify the formula for MAD(Y ) for the following distributions.
(Hint: Some of the formulas may need to be verified numerically. Find the
cdf in the appropriate section of Chapter 3. Then find the population median
MED(Y ) = M. The following trick can be used except for part c). If the
distribution is symmetric, find U = y0.75. Then D = MAD(Y ) = U − M.)
a) Cauchy (µ, σ).
b) Double exponential (θ, λ). CONTINUED
CHAPTER 3. SOME USEFUL DISTRIBUTIONS 94
c) Exponential (λ).
d) Logistic (µ, σ).
e) Normal (µ, σ 2).
f) Uniform (θ1, θ2 ).
3.4. Verify the formula for the expected value E(Y ) for the following
distributions.
a) Binomial (k, ρ).
b) Double exponential (θ, λ).
c) Exponential (λ).
d) gamma (ν, λ).
e) Logistic (µ, σ). (Hint from deCani and Stine (1986): Let Y = [µ + σW ] so
E(Y ) = µ + σE(W ) where W ∼ L(0, 1). Hence
∞
ey
E(W ) = y y 2
dy.
−∞ [1 + e ]
Now
uw|10 = [v log(v) + (1 − v) log(1 − v)] w|10 = 0
since
lim v log(v) = 0.
v→0
Now
1 1 1
log(v) log(1 − v)
− udw = − dv − dv = 2π 2/6 = π 2 /3
0 0 1−v 0 v
using
1 1
log(v) log(1 − v)
dv = dv = −π 2/6.)
0 1−v 0 v
α = P [Y ≤ Gα ].
Using
1 1
Y 1/3 ≈ N((νλ)1/3 (1 − ), (νλ)2/3 ),
9ν 9ν
show that
1 1
Gα ≈ νλ[zα + 1 − ]3
9ν 9ν
where zα is the standard normal percentile, α = Φ(zα ).
3.7. Suppose that Y1 , ..., Yn are iid from a power (λ) distribution. Suggest
a robust estimator for λ
a) based on Yi and
b) based on Wi = − log(Yi ).
3.8. Suppose that Y1 , ..., Yn are iid from a truncated extreme value
TEV(λ) distribution. Find a robust estimator for λ
a) based on Yi and
b) based on Wi = eYi − 1.
3.9. Other parameterizations for the Rayleigh distribution are possible.
For example, take µ = 0 and λ = 2σ 2. Then W is Rayleigh RAY(λ), if the
pdf of W is
2w
f(w) = exp(−w2 /λ)
λ
where λ and w are both positive.
The cdf of W is F (w) = 1 − exp(−w2/λ) for w > 0.
E(W ) = λ1/2 Γ(1 + 1/2).
VAR(W ) = λΓ(2) − (E(W ))2.
r
E(W r ) = λr/2 Γ(1 + ) for r > −2.
2
MED(W ) = λ log(2).
W is RAY(λ) if W is Weibull W (λ, 2). Thus W 2 ∼ EXP(λ). If all wi > 0,
then a trimming rule is keep wi if 0 ≤ wi ≤ 3.0(1 + 2/n)MED(n).
a) Find the median MED(W ).
CHAPTER 3. SOME USEFUL DISTRIBUTIONS 97
exp[(y − µ)/σ] − 1
F (y) =
1 + exp[(y − µ)/σ)]
Truncated Distributions
100
CHAPTER 4. TRUNCATED DISTRIBUTIONS 101
and variance
d) If c = d then
2
σW (a, b) = (β − α)σT2 (a, b) + [α − α2 + 1 − β − (1 − β)2 + 2α(1 − β)]d2.
CHAPTER 4. TRUNCATED DISTRIBUTIONS 103
Proof. We will prove b) since its proof contains the most algebra. Now
2
σW = α(µT − c)2 + (β − α)(σT2 + µ2T ) + (1 − β)(µT + d)2 − µ2W .
and
2 2 1 − 12 (k 2 + 2k + 2)e−k
b) E(Y ) = 2λ .
1 − e−k
See Problem 4.9 for a related result.
Proof. a) Note that
kλ
y −y/λ
ck E(Y ) = e dy
0 λ
kλ
−y/λ kλ
= −ye |0 + e−y/λ dy
0
(use integration by parts). So ck E(Y ) =
b) Note that
2
kλ
y 2 −y/λ
ck E(Y ) = e dy.
0 λ
Since
d
[−(y 2 + 2λy + 2λ2 )e−y/λ ]
dy
1 −y/λ 2
= e (y + 2λy + 2λ2 ) − e−y/λ (2y + 2λ)
λ
1
= y 2 e−y/λ ,
λ
2
we have ck E(Y ) =
[−(y 2 + 2λy + 2λ2 )e−y/λ ]kλ
0
for a ≤ y ≤ b.
Lemma 4.4. a) E(Y ) = µ.
2 1 − 12 (c2 + 2c + 2)e−c
b) VAR(Y ) = 2λ .
1 − e−c
where φ is the standard normal pdf and Φ is the standard normal cdf. The
indicator function
I[a,b](y) = 1 if a ≤ y ≤ b
and is zero otherwise.
CHAPTER 4. TRUNCATED DISTRIBUTIONS 106
Lemma 4.5.
φ( a−µ
σ
) − φ( b−µ
σ
)
E(Y ) = µ + σ,
Φ( b−µ
σ
) − Φ( a−µ
σ
)
and VAR(Y ) =
2
( a−µ
)φ( a−µ
) − ( b−µ
)φ( b−µ
) φ( a−µ
) − φ( b−µ
)
σ2 1 + σ σ σ σ
− σ2 σ σ
.
Φ( b−µ
σ
) − Φ( a−µ
σ
) Φ( b−µ
σ
) − Φ( a−µ
σ
)
σ
2
µ µ 2 a−µ a−µ b−µ b−µ 1
= 2 E(Y ) − +σ ( )φ( )−( )φ( )+ .
c c σ σ σ σ c
Using
1
VAR(Y ) = c E(Y 2 ) − (E(Y ))2
c
gives the result. QED
Corollary 4.6. Let Y be T N(µ, σ 2 , a = µ − kσ, b = µ + kσ). Then
E(Y ) = µ and VAR(Y ) =
2 2kφ(k)
σ 1− .
2Φ(k) − 1
CHAPTER 4. TRUNCATED DISTRIBUTIONS 108
k VAR(Y )
2.0 0.774σ 2
2.5 0.911σ 2
3.0 0.973σ 2
3.5 0.994σ 2
4.0 0.999σ 2
Proof. Use the symmetry of φ, the fact that Φ(−x) = 1 − Φ(x), and the
above lemma to get the result. QED
Examining VAR(Y ) for several values of k shows that the T N(µ, σ 2, a =
µ − kσ, b = µ + kσ) distribution does not change much for k > 3.0. See Table
4.1.
See, for example, Bickel (1965). This formula is useful since the variance of
the truncated distribution σT2 (a, b) has been computed for several distribu-
tions in the previous sections.
Definition 4.4. An estimator Dn is a location and scale equivariant
estimator if
then
√ √
n[Dn (X) − µD (FX )] = n[α + βDn (Y ) − (α + βµD (FY ))] → N(0, β 2 σD
2
D
).
where α = Φ(−z), and z = kΦ−1 (0.75). For the two stage estimators, round
100α up to the nearest integer J. Then use αJ = J/100 and zJ = −Φ−1 (αJ )
in Equation (4.4).
Proof. If Y follows the normal N(µ, σ 2) distribution, then a = µ −
kMAD(Y ) and b = µ+kMAD(Y ) where MAD(Y ) = Φ−1 (0.75)σ. It is enough
CHAPTER 4. TRUNCATED DISTRIBUTIONS 112
4.6 Simulation
In statistics, simulation uses computer generated pseudo-random variables in
place of real data. This artificial data can be used just like real data to pro-
duce histograms and confidence intervals and to compare estimators. Since
the artificial data is under the investigator’s control, often the theoretical
behavior of the statistic is known. This knowledge can be used to estimate
population quantities (such as MAD(Y )) that are otherwise hard to compute
and to check whether software is running correctly.
Example 4.3. The R/Splus software is especially useful for generating
random variables. The command
Y <- rnorm(100)
creates a vector Y that contains 100 pseudo iid N(0,1) variables. More gen-
erally, the command
Y <- rnorm(100,10,sd=4)
creates a vector Y that contains 100 pseudo iid N(10, 16) variables since
42 = 16. To study the sampling distribution of Y n , we could generate K
N(0, 1) samples of size n, and compute Y n,1 , ..., Y n,K where the notation
Y n,j denotes the sample mean of the n pseudo-variates from the jth sample.
The command
M <- matrix(rnorm(1000),nrow=100,ncol=10)
creates a 100 × 10 matrix containing 100 samples of size 10. (Note that
100(10) = 1000.) The command
creates the vector M100 of length 100 which contains Y n,1 , ..., Y n,K where
K = 100 and n = 10. A histogram from this vector should resemble the pdf
of a N(0, 0.1) random variable. The sample mean and variance of the 100
vector entries should be close to 0 and 0.1, respectively.
Example 4.4. Similarly the commands
gets the sample mean for each (row) sample of 10 observations. The command
M <- matrix(rexp(10000),nrow=100,ncol=100)
creates a 100 × 100 matrix containing 100 samples of size 100 exponential(1)
(pseudo) variates. (Note that 100(100) = 10000.) The command
gets the sample mean for each (row) sample of 100 observations. The com-
mands
will make histograms of the 100 sample means. The first histogram should
be more skewed than the second, illustrating the central limit theorem.
Example 4.5. As a slightly more complicated example, suppose that
it is desired to approximate the value of MAD(Y ) when Y is the mixture
distribution with cdf F (y) = 0.95Φ(y) + 0.05Φ(y/3). That is, roughly 95% of
the variates come from a N(0, 1) distribution and 5% from a N(0, 9) distribu-
tion. Since MAD(n) is a good estimator of MAD(Y ), the following R/Splus
commands can be used to approximate MAD(Y ).
0.95*pnorm(.7) + 0.05*pnorm(.7/3)
which gives the value 0.749747. Hence the approximation was quite good.
Definition 4.5. Let T1,n and T2,n be two estimators of a parameter τ
such that
nδ (T1,n − τ ) → N(0, σ12(F ))
D
CHAPTER 4. TRUNCATED DISTRIBUTIONS 115
F n Y MED(n) 1% TM TS,n
N(0,1) 10 1.116 1.454 1.116 1.166
N(0,1) 50 0.973 1.556 0.973 0.974
N(0,1) 100 1.040 1.625 1.048 1.044
N(0,1) 1000 1.006 1.558 1.008 1.010
N(0,1) ∞ 1.000 1.571 1.004 1.004
DE(0,1) 10 1.919 1.403 1.919 1.646
DE(0,1) 50 2.003 1.400 2.003 1.777
DE(0,1) 100 1.894 0.979 1.766 1.595
DE(0,1) 1000 2.080 1.056 1.977 1.886
DE(0,1) ∞ 2.000 1.000 1.878 1.804
and
nδ (T2,n − τ ) → N(0, σ22 (F )),
D
σ22 (F ) AV (T2,n )
AE(T1,n, T2,n ) = = .
σ12 (F ) AV (T1,n )
This definition brings up several issues. First, both estimators must have
the same convergence rate nδ . Usually δ = 0.5. If Ti,n has convergence rate
nδi , then estimator T1,n is judged to be better than T2,n if δ1 > δ2. Secondly,
the two estimators need to estimate the same parameter τ. This condition
will often not hold unless the distribution is symmetric about µ. Then τ = µ
is a natural choice. Thirdly, robust estimators are often judged by their
Gaussian efficiency with respect to the sample mean (thus F is the normal
distribution). Since the normal distribution is a location–scale family, it is
often enough to compute the AE for the standard normal distribution. If the
data come from a distribution F and the AE can be computed, then T1,n is
judged to be a better estimator at the data than T2,n if the AE > 1.
In simulation studies, typically the underlying distribution F belongs to a
symmetric location–scale family. There are at least two reasons for using such
distributions. First, if the distribution is symmetric, then the population
CHAPTER 4. TRUNCATED DISTRIBUTIONS 116
in the rows n = ∞. The simulations were performed for normal and double
exponential data, and the simulated values are close to the theoretical values.
In order for a location estimator to be used for inference, there must exist
a useful SE and a useful cutoff value td where the degrees of freedom d is
a function of n. Two criteria will be used to evaluate the CI’s. First, the
observed coverage is the proportion of the K = 500 runs for which the CI
contained the parameter estimated by Dn . This proportion should be near
the nominal coverage 0.95. Notice that if W is the proportion of runs where
the CI contains the parameter,
then KW is a binomial random variable.
Hence the SE of W is p̂(1 − p̂)/K ≈ 0.013 for the observed proportion
CHAPTER 4. TRUNCATED DISTRIBUTIONS 118
p̂ ∈ [0.9, 0.95], and an observed coverage between 0.92 and 0.98 suggests that
the observed coverage is close to the nominal coverage of √ 0.95.
The second criterion is the scaled length of the CI = n CI length =
√
n(2)(td,0.975)(SE(Dn )) ≈ 2(1.96)(σD )
√ D 2
where the approximation holds if d >√30, if n(Dn − µD ) → N(0, σD ), and
if SE(Dn ) is a good estimator of σD / n for the given value of n.
Tables 4.3 and 4.4 can be used to examine the six different interval esti-
mators. A good estimator should have an observed coverage p̂ ∈ [.92, .98],
and a small scaled length. In Table 4.3, coverages were good for N(0, 1)
data, except the interval (v) where SERM (Ln , Un ) is slightly too small for
CHAPTER 4. TRUNCATED DISTRIBUTIONS 119
n ≤ 100. The coverages for the C(0,1) and DE(0,1) data were all good even
for n = 10.
For the mixture 0.75N(0, 1) + 0.25N(100, 1), the “coverage” counted the
number of times 0 was contained in the interval and divided the result by 500.
These rows do not give a genuine coverage since the parameter µD estimated
by Dn is not 0 for any of these estimators. For example Y estimates µ = 25.
Since the median, 25% trimmed mean, and TS,n trim the same proportion of
cases to the left as to the right, MED(n) is estimating MED(Y ) ≈ Φ−1 (2/3) ≈
0.43 while the parameter estimated by TS,n is approximately the mean of a
truncated standard normal random variable where the truncation points are
Φ−1 (.25) and ∞. The 25% trimmed mean also has trouble since the number
of outliers is a binomial(n, 0.25) random variable. Hence approximately half
of the samples have more than 25% outliers and approximately half of the
samples have less than 25% outliers. This fact causes the 25% trimmed mean
to have great variability. The parameter estimated by TA,n is zero to several
decimal places. Hence the coverage of the TA,n interval is quite high.
The exponential(1) distribution is skewed, so the central limit theorem
is not a good approximation for n = 10. The estimators Y , TA,n , TS,n , MED(n)
and the 25% trimmed mean are estimating the parameters 1, 0.89155, 0.83071,
log(2) and 0.73838 respectively. Now the coverages of TA,n and TS,n are
slightly too small. For example, TS,n is asymptotically equivalent to the 10%
trimmed mean since the metrically trimmed mean truncates the largest 9.3%
of the cases, asymptotically. For small n, the trimming proportion will be
quite variable and the mean of a truncated exponential distribution with
the largest γ percent of cases trimmed varies with γ. This variability of the
truncated mean does not occur for symmetric distributions if the trimming
is symmetric since then the truncated mean µT is the point of symmetry
regardless of the amount of truncation.
Examining Table 4.4 for N(0,1) data shows that the scaled lengths of the
first 3 intervals are about the√same. The rows labeled ∞ give the scaled
length 2(1.96)(σD ) expected if nSE is a good estimator of σD . The median
interval and 25% trimmed mean interval are noticeably
√ larger than the clas-
sical interval. Since the degrees of freedom d ≈ n for the median intervals,
td,0.975 is considerably larger than 1.96 = z0.975 for n ≤ 100.
The intervals for the C(0,1) and DE(0,1) data behave about as expected.
The classical interval is very long at C(0,1) data since the first moment of
C(0,1) data does not exist. Notice that for n ≥ 50, all of the resistant
intervals are shorter on average than the classical intervals for DE(0,1) data.
CHAPTER 4. TRUNCATED DISTRIBUTIONS 121
For the mixture distribution, examining the length of the interval should
be fairer than examining the “coverage.” The length of the 25% trimmed
mean is long since about half of the time the trimmed data contains no
outliers while half of the time the trimmed data does contain outliers. When
n = 100, the length of the TS,n interval is quite long. This occurs because
the TS,n will usually trim all outliers, but the actual proportion of outliers
is binomial(100, 0.25). Hence TS,n is sometimes the 20% trimmed mean and
sometimes the 30% trimmed mean. But the parameter µT estimated by the
γ % trimmed mean varies quite a bit with γ. When n = 1000, the trimming
proportion is much less variable, and the CI length is shorter.
For exponential(1) data, 2(1.96)(σD ) = 3.9199 for Y and MED(n). The
25% trimmed mean appears to be the best of the six intervals since the scaled
length is the smallest while the coverage is good.
4.7 Complements
Several points about resistant location estimators need to be made. First,
by far the most important step in analyzing location data is to
check whether outliers are present with a plot of the data. Secondly,
no single procedure will dominate all other procedures. In particular, it is
unlikely that the sample mean will be replaced by a robust estimator. The
sample mean works very well for distributions with second moments if the
second moment is small. In particular, the sample mean works well for many
skewed and discrete distributions. Thirdly, the mean and the median should
usually both be computed. If a CI is needed and the data is thought to be
symmetric, several resistant CI’s should be computed and compared with the
classical interval. Fourthly, in order to perform hypothesis testing, plausible
values for the unknown parameter must be given. The mean and median of
the population are fairly simple parameters even if the population is skewed
while the truncated population mean is considerably more complex.
With some robust estimators, it very difficult to determine what the es-
timator is estimating if the population is not symmetric. In particular, the
difficulty in finding plausible values of the population quantities estimated
by M, L, and R estimators may be one reason why these estimators are not
widely used. For testing hypotheses, the following population quantities are
listed in order of increasing complexity.
1. The population median MED(Y ).
CHAPTER 4. TRUNCATED DISTRIBUTIONS 122
Bickel (1965), Prescott (1978), and Olive (2001) give formulas similar to
Equations (4.4) and (4.5). Gross (1976), Guenther (1969) and Lax (1985) are
useful references for confidence intervals. Andrews, Bickel, Hampel, Huber,
Rogers and Tukey (1972) is a very well known simulation study for robust
location estimators.
In Section 4.6, only intervals that are simple to compute by hand for
sample sizes of ten or so were considered. The interval based on MED(n)
(see Application 2.2 and the column “MED” in Tables 4.3 and 4.4) is even
easier to compute than the classical interval, kept its coverage pretty well,
and was frequently shorter than the classical interval.
Stigler (1973a) showed that the trimmed mean has a limiting normal
distribution even if the population is discrete provided that the asymptotic
truncation points a and b have zero probability; however, in finite samples
the trimmed mean can perform poorly if there are gaps in the distribution
near the trimming proportions.
The estimators TS,n and TA,n depend on a parameter k. Smaller values of
k should have smaller CI lengths if the data has heavy tails while larger values
of k should perform better for light tailed distributions. In simulations, TS,n
performed well for k > 1, but the variability of TA,n was too large for n ≤ 100
for Gaussian data if 1 < k < 5. These estimators also depend on the grid
C of trimming proportions. Using C = {0, 0.01, 0.02, ..., 0.49, 0.5} makes the
estimators easy to compute, but TS,n will perform better if the much coarser
grid Cc = {0, 0.01, 0.10, 0.25, 0.40, 0.49, 0.5} is used. The performance does
not change much for symmetric data, but can improve considerably if the
data is skewed. The estimator can still perform rather poorly if the data is
asymmetric and the trimming proportion of the metrically trimmed mean is
near one of these allowed trimming proportions. For example if k = 3.5 and
the data is exponential(1), the metrically trimmed mean trims approximately
9.3% of the cases. Hence the TS,n is often the 25% and the 10% trimmed
CHAPTER 4. TRUNCATED DISTRIBUTIONS 123
mean for small n. When k = 4.5, TS,n with grid Cc is usually the 10%
trimmed mean and hence performs well on exponential(1) data.
TA,n is the estimator most like high breakdown M–estimators proposed
in the literature. These estimators basically use a random amount of trim-
ming and work well on symmetric data. Estimators that give zero weight to
distant outliers (“hard rejection”) can work well on “contaminated normal”
populations such as (1 −
)N(0, 1) +
N(µs , 1). Of course
∈ (0, 0.5) and µs
can always be chosen so that these estimators perform poorly. Stigler (1977)
argues that complicated robust estimators are not needed.
4.8 Problems
PROBLEMS WITH AN ASTERISK * ARE ESPECIALLY USE-
FUL.
4.1∗. Suppose the random variable X has cdf FX (x) = 0.9 Φ(x − 10) +
0.1 FW (x) where Φ(x − 10) is the cdf of a normal N(10, 1) random variable
with mean 10 and variance 1 and FW (x) is the cdf of the random variable
W that satisfies P (W = 200) = 1.
a) Find E(W ).
b) Find E(X).
4.2. Suppose the random variable X has cdf FX (x) = 0.9 FZ (x) +
0.1 FW (x) where FZ is the cdf of a gamma(α = 10, β = 1) random variable
with mean 10 and variance 10 and FW (x) is the cdf of the random variable
W that satisfies P (W = 400) = 1.
a) Find E(W ).
b) Find E(X).
4.3. a) Prove Lemma 4.2 a).
b) Prove Lemma 4.2 c).
c) Prove Lemma 4.2 d).
d) Prove Lemma 4.2 e).
4.4. Suppose that F is the cdf from a distribution that is symmetric
about 0. Suppose a = −b and α = F (a) = 1 − β = 1 − F (b). Show that
2
σW (a, b) σT2 (a, b) 2α(F −1 (α))2
= + .
(β − α)2 1 − 2α (1 − 2α)2
CHAPTER 4. TRUNCATED DISTRIBUTIONS 124
n
4.5. Recall
n that L(Mn ) = i=1 I[Yi < MED(n) − k MAD(n)] and n −
U(Mn ) = i=1 I[Yi > MED(n) + k MAD(n)] where the indicator variable
I(A) = 1 if event A occurs and is zero otherwise. Show that TS,n is a
randomly trimmed mean. (Hint: round
up to the nearest integer, say Jn . Then TS,n is the Jn % trimmed mean with
Ln = (Jn /100) n
and Un = n − Ln .)
4.6. Show that TA,n is a randomly trimmed mean. (Hint: To get Ln ,
round 100L(Mn )/n up to the nearest integer Jn . Then Ln = (Jn /100) n
.
Round 100[n − U(Mn )]/n up to the nearest integer Kn . Then Un = (100 −
Kn )n/100
.)
4.7∗. Let F be the N(0, 1) cdf. Show that the efficiency of the sample
median MED(n) with respect to the sample mean Y n is AE ≈ 0.64.
4.8∗. Let F be the DE(0, 1) cdf. Show that the efficiency of the sample
median MED(n) with respect to the sample mean Y n is AE ≈ 2.0.
4.9. If Y is T EXP (λ, b = kλ) for k > 0, show that a)
k
E(Y ) = λ 1 − k .
e −1
b)
2 2 (0.5k 2 + k)
E(Y ) = 2λ 1 − .
ek − 1
R/Splus problems
Warning: Use the command source(“A:/rpack.txt”) to download
the programs. See Preface or Section 14.2. Typing the name of the
rpack function, eg rcisim, will display the code for the function. Use the
args command, eg args(rcisim), to display the needed arguments for the
function.
4.10. a) Download the R/Splus function nav that computes Equation
(4.4) from Lemma 4.8.
CHAPTER 4. TRUNCATED DISTRIBUTIONS 125
b) Find the asymptotic variance of the α trimmed mean for α = 0.01, 0.1,
0.25 and 0.49.
c) Find the asymptotic variance of TA,n for for k = 2, 3, 4, 5 and 6.
4.11. a) Download the R/Splus function deav that computes Equation
(4.5) from Lemma 4.9.
b) Find the asymptotic variance of the α trimmed mean for α = 0.01, 0.1,
0.25 and 0.49.
c) Find the asymptotic variance of TA,n for for k = 2, 3, 4, 5 and 6.
4.12. a) Download the R/Splus function cav that finds n AV for the
Cauchy(0,1) distribution.
b) Find the asymptotic variance of the α trimmed mean for α = 0.01, 0.1,
0.25 and 0.49.
c) Find the asymptotic variance of TA,n for for k = 2, 3, 4, 5 and 6.
4.13. a) Download the R/Splus function rcisim to reproduce Tables
4.3 and 4.4. Two lines need to be changed with each CI. One line is the
output line that calls the CI and the other line is the parameter estimated
for exponential(1) data. The program below is for the classical interval.
Thus the program calls the function cci used in Problem 2.16. The functions
medci, tmci, atmci, stmci, med2ci, cgci and bg2ci given in Problems 2.22
– 2.28 are also interesting.
b) Enter the following commands, obtain the output and explain what
the output shows.
i) rcisim(n,type=1) for n = 10, 50, 100
ii) rcisim(n,type=2) for n = 10, 50, 100
iii) rcisim(n,type=3) for n = 10, 50, 100
iv) rcisim(n,type=4) for n = 10, 50, 100
v) rcisim(n,type=5) for n = 10, 50, 100
Y = Xβ + e, (5.2)
126
CHAPTER 5. MULTIPLE LINEAR REGRESSION 127
ri (b) = ri = Yi − xTi b = Yi − Ŷi . The order statistics for the absolute residuals
are denoted by
|r|(1) ≤ |r|(2) ≤ · · · ≤ |r|(n) .
Two of the most used classical regression methods are ordinary least squares
(OLS) and least absolute deviations (L1 ).
Definition 5.2. The ordinary least squares estimator β̂OLS minimizes
n
QOLS (b) = ri2 (b), (5.4)
i=1
n
QL1 (b) = |ri (b)|. (5.5)
i=1
1500
YHAT
500
-500
80
60
YHAT^(0.5)
40
20
0
8
7
YHAT^(0)
6
5
1.95
YHAT^(-0.5)
1.90
1.85
1.002
0.998
YHAT^(-1)
0.994
120
3000
100
80
Y**(0.5)
2000
Y
60
1000
40
20
0
-500 0 500 1000 1500 2000 -500 0 500 1000 1500 2000
YHAT YHAT
1.48
7
Y**(-2/3)
LOG(Y)
1.46
6
1.44
5
-500 0 500 1000 1500 2000 -500 0 500 1000 1500 2000
YHAT YHAT
the logarithm log S of the shell mass S and a constant. With this starting
point, we might expect a log transformation of M to be needed because M
and S are both mass measurements and log S is being used as a predictor.
Using log M would essentially reduce all measurements to the scale of length.
The Box–Cox likelihood method gave λ̂0 = 0.28 with approximate 95 percent
confidence interval 0.15 to 0.4. The log transformation is excluded under this
inference leading to the possibility of using different transformations of the
two mass measurements.
The FFλ plot (not shown, but very similar to Figure 5.1) exhibits strong
linear relations, the correlations ranging from 0.9716 to 0.9999. Shown in
Figure 5.3 are transformation plots of Y (λ) versus Ŷ for four values of λ. A
striking feature of these plots is the two points that stand out in three of the
four plots (cases 8 and 48). The Box–Cox estimate λ̂ = 0.28 is evidently in-
fluenced by the two outlying points and, judging deviations from the OLS line
in Figure 5.3c, the mean function for the remaining points is curved. In other
words, the Box–Cox estimate is allowing some visually evident curvature in
the bulk of the data so it can accommodate the two outlying points. Recom-
puting the estimate of λo without the highlighted points gives λ̂o = −0.02,
which is in good agreement with the log transformation anticipated at the
outset. Reconstruction of the plots of Ŷ versus Y (λ) indicated that now the
information for the transformation is consistent throughout the data on the
horizontal axis of the plot.
The essential point of this example is that observations that influence the
choice of power transformation are often easily identified in a transformation
plot of Ŷ versus Y (λ) when the FFλ subplots are strongly linear.
The easily verified assumption that there is strong linearity in the FFλ
plot is needed since if λo ∈ [−1, 1], then
for all λ ∈ [−1, 1]. Consequently, for any value of λ ∈ [−1, 1], Ŷ (λ) is essen-
tially a linear function of the fitted values Ŷ (λo) for the true λo , although we
do not know λo itself. However, to estimate λo graphically, we could select
∗
any fixed value λ∗ ∈ [−1, 1] and then plot Ŷ (λ ) versus Y (λ) for several values
of λ and find the λ ∈ Λc for which the plot is linear with constant variance.
∗
This simple graphical procedure will then work because a plot of Ŷ (λ ) versus
Y (λ) is equivalent to a plot of cλ∗ + dλ∗ Ŷ (λo ) versus Y (λ) by Equation (5.8).
Since the plot of Ŷ (1) versus Y (λ) differs from a plot of Ŷ versus Y (λ) by a
CHAPTER 5. MULTIPLE LINEAR REGRESSION 134
4
2.0
3
1.5
Y**(-0.25)
LOG(Y)
2
8
1.0
1
0.5
48 48
0.0
-10 0 10 20 30 40 -10 0 10 20 30 40
YHAT YHAT
40
Y**(0.28)
30
4
20
2
8
10
48
48 8
0
-10 0 10 20 30 40 -10 0 10 20 30 40
YHAT YHAT
so that any one set of population fitted values is an exact linear function
of any other set provided the τλ ’s are nonzero. See Cook and Olive (2001).
This result indicates that sample FFλ plots will be linear when E(w|wT η) is
linear, although Equation (5.9) does not by itself guarantee high correlations.
However, the strength of the relationship (5.8) can be checked easily by
inspecting the FFλ plot.
Secondly, if the FFλ subplots are not strongly linear, and if there is non-
linearity present in the scatterplot matrix of the nontrivial predictors, then
transforming the predictors to remove the nonlinearity will often
be a useful procedure. The linearizing of the predictor relationships could
be done by using marginal power transformations or by transforming the
joint distribution of the predictors towards an elliptically contoured distri-
bution. The linearization might also be done by using simultaneous power
transformations λ = (λ2 , . . . , λp )T of the predictors so that the vector wλ
(λ ) (λ )
= (x2 2 , ..., xp p )T of transformed predictors is approximately multivariate
normal. A method for doing this was developed by Velilla (1993). (The basic
idea is the same as that underlying the likelihood approach of Box and Cox
CHAPTER 5. MULTIPLE LINEAR REGRESSION 136
50
40
30
YHAT
20
10
0
8 10
YHAT^(0.5)
6
4
2
3.5
YHAT^(0)
2.5
1.5
1.8
1.4
YHAT^(-0.5)
1.0
1.0
0.9
YHAT^(-1)
0.8
0.7
0 10 20 30 40 50 1.5 2.5 3.5 0.7 0.8 0.9 1.0
Figure 5.4: FFλ Plot for Mussel Data with Original Predictors
300
250
length
200
150
60
50
width
40
30
20
140
height
100
80
300
200
shell
100
0
300
250
length
200
150
4.2
3.8
Log W
3.4
3.0
140
height
100
80
6
5
Log S
4
3
where xI/S denotes the predictors in I that are not in S. Since this is true
regardless of the values of the predictors, β O = 0 and the sample correlation
corr(βT xi , βTI xI,i) = 1.0 for the population model if S ⊆ I.
This section proposes a graphical method for evaluating candidate sub-
models. Let β̂ be the estimate of β obtained from the regression of Y on all
of the terms x. Denote the residuals and fitted values from the full model by
T T
ri = Yi − β̂ xi = Yi − Ŷi and Ŷi = β̂ xi respectively. Similarly, let β̂ I be the
CHAPTER 5. MULTIPLE LINEAR REGRESSION 139
visual aids. The subset I is good if the plotted points cluster tightly about
the identity line in both plots. In particular, the OLS line and the identity
line should nearly coincide near the origin in the RR plot.
To verify that the six plots are useful for assessing variable selection,
the following notation will be useful. Suppose that all submodels include
a constant and that X is the full rank n × p design matrix for the full
model. Let the corresponding vectors of OLS fitted values and residuals be
Ŷ = X(X T X)−1 X T Y = HY and r = (I − H)Y , respectively. Sup-
pose that X I is the n × k design matrix for the candidate submodel and
that the corresponding vectors of OLS fitted values and residuals are Ŷ I =
X I (X TI X I )−1 X TI Y = H I Y and r I = (I − H I )Y , respectively. For mul-
tiple linear regression, recall that if the candidate model of xI has k terms
(including the constant), then the FI statistic for testing whether the p − k
predictor variables in xO can be deleted is
SSE(I) − SSE SSE n − p SSE(I)
FI = / = [ − 1]
(n − k) − (n − p) n − p p − k SSE
where SSE is the error sum of squares from the full model and SSE(I) is the
error sum of squares from the candidate submodel. Also recall that
SSE(I)
Cp (I) = + 2k − n = (p − k)(FI − 1) + k
MSE
where MSE is the error mean square for the full model. Notice that Cp (I) ≤
2k if and only if FI ≤ p/(p − k). Remark 5.3 below suggests that for subsets
I with k terms, submodels with Cp (I) ≤ 2k are especially interesting.
A plot can be very useful if the OLS line can be compared to a reference
line and if the OLS slope is related to some quantity of interest. Suppose
that a plot of w versus z places w on the horizontal axis and z on the vertical
axis. Then denote the OLS line by ẑ = a + bw. The following proposition
shows that the FF, RR and forward response plots will cluster about the
identity line. Notice that the proposition is a property of OLS and holds
even if the data does not follow an MLR model. Let corr(x, y) denote the
correlation between x and y.
Proposition 5.1. Suppose that every submodel contains a constant and
that X is a full rank matrix.
Forward Response Plot: i) If w = ŶI and z = Y then the OLS line is the
CHAPTER 5. MULTIPLE LINEAR REGRESSION 141
identity line.
ii) If w = Y and z = ŶI then the OLS line has slope b = [corr(Y, ŶI )]2 = R2I
and intercept a = Y (1 − R2I ) where Y = ni=1 Yi /n and R2I is the coefficient
of multiple determination from the candidate model.
FF Plot: iii) If w = ŶI and z = Ŷ then the OLS line is the identity line.
Note that ESP (I) = ŶI and ESP = Ŷ .
iv) If w = Ŷ and z = ŶI then the OLS line has slope b = [corr(Ŷ , ŶI )]2 =
SSR(I)/SSR and intercept a = Y [1 − (SSR(I)/SSR)] where SSR is the
regression sum of squares.
v) If w = r and z = rI then the OLS line is the identity line.
RR Plot: vi) If w = rI and z = r then a = 0 and the OLS slope b =
[corr(r, rI )]2 and
SSE n−p n−p
corr(r, rI ) = = = .
SSE(I) Cp(I) + n − 2k (p − k)FI + n − p
Also recall that the OLS line passes through the means of the two variables
(w, z).
(*) Notice that the OLS slope from regressing z on w is equal to one if
and only if the OLS slope from regressing w on z is equal to [corr(z, w)]2.
2 T
i) The slope b = 1 if ŶI,i Yi = ŶI,i . This equality holds since Ŷ I Y =
T
Y T H I Y = Y T H I H I Y = Ŷ I Ŷ I . Since b = 1, a = Y − Y = 0.
ii) By (*), the slope
2 (ŶI,i − Y )2
b = [corr(Y, ŶI )] = R2I = = SSR(I)/SST.
(Yi − Y )2
2
iii) The slope b = 1 if ŶI,i Ŷi = ŶI,i . This equality holds since
T T
Ŷ Ŷ I = Y T HH I Y = Y T H I Y = Ŷ I Ŷ I . Since b = 1, a = Y − Y = 0.
iv) From iii),
SD(Ŷ )
1= [corr(Ŷ , ŶI )].
SD(ŶI )
Hence
SD(ŶI )
corr(Ŷ , ŶI ) =
SD(Ŷ )
and the slope
SD(ŶI )
b= corr(Ŷ , ŶI ) = [corr(Ŷ , ŶI )]2 .
SD(Ŷ )
Also the slope
(ŶI,i − Y )2
b= = SSR(I)/SSR.
(Ŷi − Y )2
The result follows since a = Y − bY .
v) The OLS line passes through the origin. Hence a = 0. The slope
b = r T r I /r T r. Since r T r I = Y T (I − H)(I − H I )Y and (I − H)(I − H I ) =
I − H, the numerator r T r I = r T r and b = 1.
vi) Again a = 0 since the OLS line passes through the origin. From v),
SSE(I)
1= [corr(r, rI )].
SSE
Hence
SSE
corr(r, rI ) =
SSE(I)
and the slope
SSE
b= [corr(r, rI )] = [corr(r, rI )]2.
SSE(I)
Algebra shows that
n−p n−p
corr(r, rI ) = = . QED
Cp(I) + n − 2k (p − k)FI + n − p
CHAPTER 5. MULTIPLE LINEAR REGRESSION 143
200
100
1200
FRES
Y
0
800
-200
400
400 600 800 1000 1200 1400 400 600 800 1000 1200 1400
FFIT FFIT
200
100
1200
SRES3
0
Y
800
-200
400
400 600 800 1000 1200 1400 400 600 800 1000 1200 1400
SFIT3 SFIT3
Figure 5.7: Gladstone data: comparison of the full model and the submodel.
Remark 5.2. Note that for large n, Cp (I) < k or FI < 1 will force
corr(ESP,ESP(I)) to be high. If the estimators β̂ and β̂ I are not the OLS
estimators, the plots will be similar to the OLS plots if the correlation of the
fitted values from OLS and the alternative estimators is high (≥ 0.95).
A standard model selection procedure will often be needed to suggest
models. For example, forward selection or backward elimination could be
used. If p < 30, Furnival and Wilson (1974) provide a technique for selecting
a few candidate subsets after examining all possible subsets.
Remark 5.3. Daniel and Wood (1980, p. 85) suggest using Mallows’
graphical method for screening subsets by plotting k versus Cp (I) for models
close to or under the Cp = k line. Proposition 5.1 vi) implies that if Cp (I) ≤ k
then corr(r, rI ) and corr(ESP, ESP (I)) both go to 1.0 as n → ∞. Hence
models I that satisfy the Cp(I) ≤ k screen will contain the true model S
with high probability when n is large. This result does not guarantee that
CHAPTER 5. MULTIPLE LINEAR REGRESSION 144
200
1200
100
FRES
FFIT
0
800
600
-200
400
-200 -100 0 100 200 400 600 800 1000 1200 1400
SRES1 SFIT1
200
100
100
FRES
FRES
0
0
-200
-200
SRES2 SRES4
Figure 5.8: Gladstone data: submodels added (size)1/3, sex, age and finally
breadth.
a) RR Plot b) FF Plot
1400
200
1200
100
1000
FRES
FFIT
0
800
-100
600
-200
400
-200 -100 0 100 200 400 600 800 1000 1200 1400
SRES3 SFIT3
Figure 5.9: Gladstone data with Predictors (size)1/3, sex, and age
CHAPTER 5. MULTIPLE LINEAR REGRESSION 145
the true model S will satisfy the screen, hence overfit is likely (see Shao
1993). Let d be a lower bound on corr(r, rI ). Proposition 5.1 vi) implies that
if
1 p
Cp (I) ≤ 2k + n 2 − 1 − 2 ,
d d
then corr(r, rI ) ≥ d. The simple screen Cp (I) ≤ 2k corresponds to
p
dn ≡ 1 − .
n
To reduce the chance of overfitting, use the Cp = k line for large values of k,
but also consider models close to or under the Cp = 2k line when k ≤ p/2.
Example 5.4. The FF and RR plots can be used as a diagnostic for
whether a given numerical method is including too many variables. Glad-
stone (1905-1906) attempts to estimate the weight of the human brain (mea-
sured in grams after the death of the subject) using simple linear regression
with a variety of predictors including age in years, height in inches, head
height in mm, head length in mm, head breadth in mm, head circumference
in mm, and cephalic index. The sex (coded as 0 for females and 1 for males)
of each subject was also included. The variable cause was coded as 1 if the
cause of death was acute, 3 if the cause of death was chronic, and coded as 2
otherwise. A variable ageclass was coded as 0 if the age was under 20, 1 if the
age was between 20 and 45, and as 3 if the age was over 45. Head size, the
product of the head length, head breadth, and head height, is a volume mea-
surement, hence (size)1/3 was also used as a predictor with the same physical
dimensions as the other lengths. Thus there are 11 nontrivial predictors and
one response, and all models will also contain a constant. Nine cases were
deleted because of missing values, leaving 267 cases.
Figure 5.7 shows the forward response plots and residual plots for the full
model and the final submodel that used a constant, size1/3, age and sex.
The five cases separated from the bulk of the data in each of the four plots
correspond to five infants. These may be outliers, but the visual separation
reflects the small number of infants and toddlers in the data. A purely
numerical variable selection procedure would miss this interesting feature of
the data. We will first perform variable selection with the entire data set,
and then examine the effect of deleting the five cases. Using forward selection
and the Cp statistic on the Gladstone data suggests the subset I5 containing
a constant, (size)1/3, age, sex, breadth, and cause with Cp(I5) = 3.199. The
CHAPTER 5. MULTIPLE LINEAR REGRESSION 146
p–values for breadth and cause were 0.03 and 0.04, respectively. The subset
I4 that deletes cause has Cp (I4) = 5.374 and the p–value for breadth was 0.05.
Figure 5.8d shows the RR plot for the subset I4. Note that the correlation
of the plotted points is very high and that the OLS and identity lines nearly
coincide.
A scatterplot matrix of the predictors and response suggests that (size)1/3
might be the best single predictor. First we regressed y = brain weight on the
eleven predictors described above (plus a constant) and obtained the residuals
ri and fitted values ŷi. Next, we regressed y on the subset I containing
(size)1/3 and a constant and obtained the residuals rI,i and the fitted values
ŷI,i. Then the RR plot of rI,i versus ri , and the FF plot of ŷI,i versus ŷi were
constructed.
For this model, the correlation in the FF plot (Figure 5.8b) was very high,
but in the RR plot the OLS line did not coincide with the identity line (Figure
5.8a). Next sex was added to I, but again the OLS and identity lines did not
coincide in the RR plot (Figure 5.8c). Hence age was added to I. Figure 5.9a
shows the RR plot with the OLS and identity lines added. These two lines
now nearly coincide, suggesting that a constant plus (size)1/3, sex, and age
contains the relevant predictor information. This subset has Cp(I) = 7.372,
R2I = 0.80, and σ̂I = 74.05. The full model which used 11 predictors and a
constant has R2 = 0.81 and σ̂ = 73.58. Since the Cp criterion suggests adding
breadth and cause, the Cp criterion may be leading to an overfit.
Figure 5.9b shows the FF plot. The five cases in the southwest corner
correspond to five infants. Deleting them leads to almost the same conclu-
sions, although the full model now has R2 = 0.66 and σ̂ = 73.48 while the
submodel has R2I = 0.64 and σ̂I = 73.89.
Example 5.5. Cook and Weisberg (1999a, p. 261, 371) describe a data
set where rats were injected with a dose of a drug approximately proportional
to body weight. The data set is included as the file rat.lsp in the Arc software
and can be obtained from the website (http://www.stat.umn.edu/arc/). The
response Y is the fraction of the drug recovered from the rat’s liver. The three
predictors are the body weight of the rat, the dose of the drug, and the liver
weight. The experimenter expected the response to be independent of the
predictors, and 19 cases were used. However, the Cp criterion suggests using
the model with a constant, dose and body weight, both of whose coefficients
were statistically significant. The FF and RR plots are shown in Figure 5.10.
The identity line and OLS lines were added to the plots as visual aids. The
CHAPTER 5. MULTIPLE LINEAR REGRESSION 147
a) RR Plot b) FF plot
0.10
0.50
0.45
0.05
full$residual
ffit
0.40
0.0
0.35
-0.05
0.30
-0.10
-0.10 -0.05 0.0 0.05 0.10 0.30 0.35 0.40 0.45 0.50
sub$residual sfit
FF plot shows one outlier, the third case, that is clearly separated from the
rest of the data.
We deleted this case and again searched for submodels. The Cp statistic
is less than one for all three simple linear regression models, and the RR and
FF plots look the same for all submodels containing a constant. Figure 5.11
shows the RR plot where the residuals from the full model are plotted against
Y − Y , the residuals from the model using no nontrivial predictors. This plot
suggests that the response Y is independent of the nontrivial predictors.
The point of this example is that a subset of outlying cases can cause
numeric second-moment criteria such as Cp to find structure that does not
exist. The FF and RR plots can sometimes detect these outlying cases,
allowing the experimenter to run the analysis without the influential cases.
The example also illustrates that global numeric criteria can suggest a model
with one or more nontrivial terms when in fact the response is independent
of the predictors.
Numerical variable selection methods for MLR are very sensitive to “influ-
ential cases” such as outliers. For the MLR model, standard case diagnostics
CHAPTER 5. MULTIPLE LINEAR REGRESSION 148
RR Plot
0.10
0.05
full$residual
0.0
-0.05
-0.10
subresidual
very different from those using the full data set, a situation that should cause
concern. Warning: deleting influential cases and outliers will often
lead to better plots and summary statistics, but the cleaned data
may no longer represent the actual population. In particular, the
resulting model may be very poor for prediction.
A thorough subset selection analysis will use the RC plots in conjunction
with the more standard numeric-based algorithms. This suggests running
the numerical variable selection procedure on the entire data set and on the
“cleaned data” set with the influential cases deleted, keeping track of inter-
esting models from both data sets. For a candidate submodel I, let Cp (I, c)
denote the value of the Cp statistic for the cleaned data. The following two
examples help illustrate the procedure.
Example 5.6. Ashworth (1842) presents a data set of 99 communities
in Great Britain. The response variable y = log(population in 1841) and the
predictors are x1, x2 , x3 and a constant where x1 is log(property value in
pounds in 1692), x2 is log(property value in pounds in 1841), and x3 is the
log(percent rate of increase in value). The initial RC plot, shown in Figure
5.12a, is far from the ideal of an evenly-populated parabolic band. Cases
14 and 55 have extremely large Cook’s distances, along with the largest
residuals. After deleting these cases and refitting OLS, Figure 5.12b shows
that the RC plot is much closer to the ideal parabolic shape. If case 16 had a
residual closer to zero, then it would be a very high leverage case and would
also be deleted.
Table 5.1 shows the summary statistics of the fits of all subsets using all
cases, and following the removal of cases 14 and 55. The two sets of results
are substantially different. On the cleaned data the submodel using just x2
is the unique clear choice, with Cp (I, c) = 0.7. On the full data set however,
none of the subsets is adequate. Thus cases 14 and 55 are responsible for all
indications that predictors x1 and x3 have any useful information about y.
This is somewhat remarkable in that these two cases have perfectly ordinary
values for all three variables.
Example 5.4 (continued). Now we will apply the RC plot to the Glad-
stone data using y = weight, x1 = age, x2 = height, x3 = head height, x4 =
head length, x5 = head breadth, x6 = head circumference, x7 = cephalic index,
x8 = sex, and x9 = (size)1/3. All submodels contain a constant.
Table 5.2 shows the summary statistics of the more interesting subset
CHAPTER 5. MULTIPLE LINEAR REGRESSION 150
14
4
16
55
0.06
3
0.04
CD2
CD
0.02
1
0.0
0
-1.5 -1.0 -0.5 0.0 0.5 1.0 -1.5 -1.0 -0.5 0.0 0.5
RES RES2
c) RR Plot d) FF Plot
17
12
14 55
1.0
0.5
10
FRES
FFIT
-0.5
14
8
6 55
-1.5
-1 0 1 2 6 8 10 12
SRES SFIT
118 234
0.06
0.08
248
0.04
CD
CD
0.04
0.02
0.0
0.0
RES RES
258
0.06
0.03
0.04
0.02
CD
CD
0.02
0.01
0.0
0.0
RES RES
regressions. The smallest Cp value came from the subset x1, x5, x8 , x9, and
in this regression x5 has a t value of −2.0. Deleting a single predictor from
an adequate regression changes the Cp by approximately t2 − 2, where t
stands for that predictor’s Student’s t in the regression – as illustrated by the
increase in Cp from 4.4 to 6.3 following deletion of x5 . Analysts must choose
between the larger regression with its smaller Cp but a predictor that does
not pass the conventional screens for statistical significance, and the smaller,
more parsimonious, regression using only apparently statistically significant
predictors, but (as assessed by Cp ) possibly less accurate predictive ability.
Figure 5.13 shows a sequence of RC plots used to identify cases 118, 234,
248 and 258 as atypical, ending up with an RC plot that is a reasonably
evenly-populated parabolic band. Using the Cp criterion on the cleaned data
CHAPTER 5. MULTIPLE LINEAR REGRESSION 152
suggests the same final submodel I found earlier – that using a constant,
x1 = age, x8 = sex and x9 = size1/3.
The five cases (230, 254, 255, 256 and 257) corresponding to the five
infants were well separated from the bulk of the data and have higher leverage
than average, and so good exploratory practice would be to remove them also
to see the effect on the model fitting. The right columns of Table 5.2 reflect
making these 9 deletions. As in the full data set, the subset x1, x5 , x8, x9 gives
the smallest Cp , but x5 is of only modest statistical significance and might
reasonably be deleted to get a more parsimonious regression. What is striking
after comparing the left and right columns of Table 5.2 is that, as was the
case with the Ashworth data set, the adequate Cp values for the cleaned data
set seem substantially smaller than their full-sample counterparts: 1.2 versus
4.4, and 2.3 versus 6.3. Since these Cp for the same p are dimensionless and
comparable, this suggests that the 9 cases removed are primarily responsible
for any additional explanatory ability in the 6 unused predictors.
CHAPTER 5. MULTIPLE LINEAR REGRESSION 153
Multiple linear regression data sets with cases that influence numerical
variable selection methods are common. Table 5.3 shows results for seven
interesting data sets. The first two rows correspond to the Ashworth data in
Example 5.6, the next 2 rows correspond to the Gladstone Data in Example
5.4, and the next 2 rows correspond to the Gladstone data with the 5 infants
deleted. Rows 7 and 8 are for the Buxton (1920) data while rows 9 and 10
are for the Tremearne (1911) data. These data sets are available from the
book’s website. Results from the final two data sets are given in the last 4
rows. The last 2 rows correspond to the rat data described in Example 5.5.
Rows 11 and 12 correspond to the Ais data that comes with Arc (Cook and
Weisberg, 1999a).
The full model used p predictors, including a constant. The final sub-
model I also included a constant, and the nontrivial predictors are listed in
the second column of Table 5.3. The third column lists p, Cp(I) and Cp (I, c)
while the first column gives the set of influential cases. Two rows are pre-
sented for each data set. The second row gives the response variable and any
predictor transformations. For example, for the Gladstone data p = 10 since
there were 9 nontrivial predictors plus a constant. Only the predictor size
was transformed, and the final submodel is the one given in Example 5.4.
For the rat data, the final submodel is the one given in Example 5.5: none
of the 3 nontrivial predictors was used.
Table 5.3 and simulations suggest that if the subset I has k predictors,
then using the Cp (I) ≤ 2k screen is better than using the conventional
Cp (I) ≤ k screen. The major and ais data sets show that deleting the
influential cases may increase the Cp statistic. Thus interesting models from
the entire data set and from the clean data set should be examined.
σ 2.
The response variable Y is the variable that you want to predict while
the predictor (or independent or explanatory) variable X is the variable used
to predict the response.
A scatterplot is a plot of W versus Z with W on the horizontal axis
and Z on the vertical axis and is used to display the conditional dis-
tribution of Z given W . For SLR the scatterplot of X versus Y is often
used.
For SLR, E(Yi ) = β1 +β2Xi and the line E(Y ) = β1 +β2X is the regression
function. VAR(Yi) = σ 2 .
For SLR, the least squares n estimators b1 and b2 minimize the least
2
squares criterion Q(η1 , η2) = i=1 (Yi − η1 − η2Xi ) . For a fixed η1 and η2 , Q
is the sum of the squared vertical deviations from the line Y = η1 + η2 X.
The least squares (OLS) line is Ŷ = b1 + b2 X where
n
(Xi − X)(Yi − Y )
β̂2 ≡ b2 = i=1n
i=1 (Xi − X)
2
and β̂1 ≡ b1 = Y − b2 X.
By the chain rule,
∂Q n
= −2 (Yi − η1 − η2 Xi )
∂η1 i=1
and
d2 Q
= 2n.
dη12
Similarly,
∂Q n
= −2 Xi (Yi − η1 − η2 Xi )
∂η2 i=1
and
d2 Q n
= 2 Xi2 .
dη12 i=1
n
n
n
Xi Yi = b1 Xi + b2 Xi2 .
i=1 i=1 i=1
Response = Y
Coefficient Estimates
Label Estimate Std. Error t-value p-value
Constant b1 se(b1) t for b1 p-value for beta_1
x b2 se(b2) to = b2/se(b2) p-value for beta_2
R Squared: r^2
Sigma hat: sqrt{MSE}
Number of cases: n
Degrees of freedom: n-2
R Squared: 0.74058
Sigma hat: 83.9447
Number of cases: 267
Degrees of freedom: 265
For SLR, Ŷi = b1 + b2 Xi is called the ith fitted value (or predicted value)
for observation Yi while the ith residual is ri = Yi − Ŷi .
n
n
2
The error (residual) sum of squares SSE = (Yi − Ŷi ) = ri2 .
i=1 i=1
For SLR, the mean square error MSE = SSE/(n − 2) is an unbiased
estimator of the error variance σ 2.
Properties of the OLS line:
i) the residuals sum to zero: ni=1 ri = 0.
ii) ni=1 Yi = ni=1 Ŷi .
iii) The independent variable and residuals are uncorrelated:
n
Xi ri = 0.
i=1
iv) The fitted values and residuals are uncorrelated: ni=1 Ŷi ri = 0.
v) The least squares line passes through the point (X, Y ).
Let the p × 1 vector β = (β1 , ..., βp)T and let the p × 1 vector xi =
(1, Xi,2 , ..., Xi,p)T . Notice that Xi,1 ≡ 1 for i = 1, ..., n. Then the multiple
linear regression (MLR) model is
for i = 1, ..., n where the ei are iid with E(ei ) = 0 and VAR(ei) = σ 2 for
i = 1, ..., n. The Yi and ei are random variables while the Xi are treated
as known constants. The parameters β1 , β2, ..., βp and σ 2 are unknown
constants that need to be estimated.
In matrix notation, these n equations become
Y = Xβ + e,
The first column of X is 1, the n × 1 vector of ones. The ith case (xTi , Yi )
corresponds to the ith row xTi of X and the ith element of Y . If the ei
are iid with zero mean and variance σ 2, then regression is used to estimate
the unknown parameters β and σ 2 . (If the Xi are random variables, then
the model is conditional on the Xi ’s. Hence the Xi ’s are still treated as
constants.)
The normal MLR model adds the assumption that the ei are iid N(0, σ 2).
That is, the error distribution in normal with zero mean and constant vari-
ance σ 2. Simple linear regression is a special case with p = 2.
The response variable Y is the variable that you want to predict while
the predictor (or independent or explanatory) variables X1 , X2 , ..., Xp are the
variables used to predict the response. Since X1 ≡ 1, sometimes X2 , ..., Xp
are called the predictor variables.
For MLR, E(Yi ) = β1 + β2Xi,2 + · · · + βpXi,p = xTi β and the hyperplane
E(Y ) = β1 + β2X2 + · · · + βpXp = xT β is the regression function. VAR(Yi) =
σ 2.
The least squares
estimators b1 , b2, ..., bp minimize theleast squares
criterion Q(η) = ni=1 (Yi − η1 − η2Xi,2 − · · · − ηp Xi,p )2 = ni=1 ri2 . For a
fixed η, Q is the sum of the squared vertical deviations from the hyperplane
H = η1 + η2X2 + · · · + ηp Xp .
The least squares estimator β̂ = b satisfies the MLR normal equations
X T Xb = X T Y
β̂ = b = (X T X)−1 X T Y .
Response = Y
Coefficient Estimates
Source df SS MS F p-value
Regression p-1 SSR MSR Fo=MSR/MSE for Ho:
Residual n-p SSE MSE β2 = · · · = βp = 0
Response = brnweight
Coefficient Estimates
Label Estimate Std. Error t-value p-value
Constant 99.8495 171.619 0.582 0.5612
size 0.220942 0.0357902 6.173 0.0000
sex 22.5491 11.2372 2.007 0.0458
breadth -1.24638 1.51386 -0.823 0.4111
circum 1.02552 0.471868 2.173 0.0307
R Squared: 0.749755
Sigma hat: 82.9175
Number of cases: 267
Degrees of freedom: 262
Know the meaning of the least squares multiple linear regression output.
Shown on the previous page is an actual Arc output and an output only using
symbols.
The 100 (1 − α) % CI for βk is bk ± t1−α/2,n−p se(bk ). If ν = n − p > 30,
use the N(0,1) cutoff z1−α/2. The corresponding 4 step t–test of hypotheses
has the following steps:
i) State the hypotheses Ho: βk = 0 Ha: βk = 0.
ii) Find the test statistic to,k = bk /se(bk ) or obtain it from output.
iii) Find the p–value from output or use the t–table: p–value =
2P (tn−p < −|to,k |).
Use the normal table or ν = ∞ in the t–table if the degrees of freedom
ν = n − p > 30.
iv) State whether you reject Ho or fail to reject Ho and give a nontechnical
sentence restating your conclusion in terms of the story problem.
Recall that Ho is rejected if the p–value < α. As a benchmark for this
textbook, use α = 0.05 if α is not given. If Ho is rejected, then conclude that
Xk is needed in the MLR model for Y given that the other p − 2 nontrivial
predictors are in the model. If you fail to reject Ho, then conclude that Xk
is not needed in the MLR model for Y given that the other p − 2 nontrivial
predictors are in the model. Note that Xk could be a very useful individual
predictor, but may not be needed if other predictors are added to the model.
It is better to use the output to get the test statistic and p–value than to use
formulas and the t–table, but exams may not give the relevant output.
Be able to perform the 4 step ANOVA F test of hypotheses:
i) State the hypotheses Ho: β2 = · · · = βp = 0 Ha: not Ho
ii) Find the test statistic F o = MSR/MSE or obtain it from output.
iii) Find the p–value from output or use the F–table: p–value =
P (Fp−1,n−p > Fo).
iv) State whether you reject Ho or fail to reject Ho. If Ho is rejected, conclude
that there is a MLR relationship between Y and the predictors X2 , ..., Xp . If
you fail to reject Ho, conclude that there is a not a MLR relationship between
Y and the predictors X2 , ..., Xp .
ii) the 100 (1 − α)% CI for E(Yh ) = xTh β = E(Ŷh ). This interval is
Ŷh ± t1−α/2,n−p se(Ŷh ). Generally se(Ŷh ) will come from output.
Suppose you want to predict a new observation Yh where Yh is indepen-
dent of Y1 , ..., Yn. Be able to find
i) the point estimator Ŷh = xTh b and the
ii) the 100 (1 − α)% prediction interval (PI) for Yh . This interval is
Ŷh ± t1−α/2,n−p se(pred). Generally se(pred) will come from output. Note that
Yh is a random variable not a parameter.
Full model
Reduced model
4 steps:
i) Ho: the reduced model is good Ha: use the full model
ii) Fo ≡ FR =
SSE(R) − SSE(F )
/MSE(F )
dfR − dfF
iii) p–value = P(FdfR −dfF ,dfF > F o). (Here dfR − dfF = p − q = number of
parameters set to 0, and dfF = n − p).
iv) Reject Ho if the p–value < α and conclude that the full model should be
used. Otherwise, fail to reject Ho and conclude that the reduced model is
good.
Given two of SSTO = ni=1 (Yi − Y )2 , SSE = ni=1 (Yi − Ŷi )2 = ni=1 ri2 ,
and SSR = ni=1 (Ŷi − Y )2, find the other sum of squares using the formula
SSTO = SSE + SSR.
Be able to find R2 = SSR/SST O = (sample correlation of Yi and Ŷi )2.
Know i) that the covariance matrix of a random vector Y is Cov(Y ) =
E[(Y − E(Y ))(Y − E(Y ))T ].
ii) E(AY ) = AE(Y ).
iii) Cov(AY ) = ACov(Y )AT .
Given the least squares model Y = Xβ + e, be able to show that
i) E(b) = β and
ii) Cov(b) = σ 2(X T X)−1 .
A matrix A is idempotent if AA = A.
An added variable plot (also called a partial regression plot) is used to
give information about the test Ho : βi = 0. The points in the plot cluster
about a line with slope = bi . If there is strong trend then Xi is needed in
the MLR for Y given that the other predictors X2 , ..., Xi−1, Xi+1 , ..., Xp are
in the model. If there is almost no trend, then Xi may not be needed in the
MLR for Y given that the other predictors X2 , ..., Xi−1, Xi+1 , ..., Xp are in
the model.
The forward response plot of ŷi versus y is used to check whether
the MLR model is appropriate. If the MLR model is appropriate, then the
plotted points should cluster about the identity line. The squared correlation
[corr(yi, ŷi )]2 = R2 . Hence the clustering is tight if R2 ≈ 1. If outliers are
present or if the plot is not linear, then the current model or data need to
be changed or corrected. Know how to decide whether the MLR model is
CHAPTER 5. MULTIPLE LINEAR REGRESSION 162
several matrices may need to be made. Each one should include Z.) Remove
or correct any gross outliers. It is often a good idea to transform the Wi
to remove any strong nonlinearities from the predictors. Eventually
you will find a response variable Y = tZ (Z) and nontrivial predictor variables
X2 , ..., Xp for the full model. Interactions such as Xk = Wi Wj and powers
such as Xk = Wi2 may be of interest. Indicator variables are often used
in interactions, but do not transform an indicator variable. The forward
response plot for the full model should be linear and the residual plot should
be ellipsoidal with zero trend. Find the OLS output. The statistic R2 gives
the proportion of the variance of Y explained by the predictors and is of
some importance.
Variable selection is closely related to the change in F test. You are
seeking a subset I of the variables to keep in the model. The submodel I
will always contain a constant and will have k − 1 nontrivial predictors where
1 ≤ k ≤ p. Know how to find candidate submodels from output.
Forward selection starts with a constant = W1 . Step 1) k = 2: compute
Cp for all models containing the constant and a single predictor Xi . Keep
the predictor W2 = Xj , say, that corresponds to the model with the smallest
value of Cp .
Step 2) k = 3: Fit all models with k = 3 that contain W1 and W2 . Keep the
predictor W3 that minimizes Cp . ...
Step j) k = j + 1: Fit all models with k = j + 1 that contains W1 , W2 , ..., Wj .
Keep the predictor Wj+1 that minimizes Cp . ...
Step p − 1): Fit the full model.
Backward elimination: All models contain a constant = U1 . Step 1)
k = p: Start with the full model that contains X1 , ..., Xp. We will also say
that the full model contains U1 , ..., Up where U1 = X1 but Ui need not equal
Xi for i > 1.
Step 2) k = p − 1: fit each model with p − 1 predictors including a constant.
Delete the predictor Up , say, that corresponds to the model with the smallest
Cp . Keep U1 , ..., Up−1.
Step 3) k = p− 2: fit each model with p− 2 predictors and a constant. Delete
the predictor Up−1 that corresponds to the smallest Cp . Keep U1 , ..., Up−2. ...
Step j) k = p − j + 1: fit each model with p − j + 1 predictors and a
constant. Delete the predictor Up−j+2 that corresponds to the smallest Cp.
Keep U1, ..., Up−j+1. ...
CHAPTER 5. MULTIPLE LINEAR REGRESSION 165
5.4 Complements
Algorithms for OLS are described in Datta (1995), Dongarra, Moler, Bunch
and Stewart (1979), and Golub and Van Loan (1989). Algorithms for L1
are described in Adcock and Meade (1997), Barrodale and Roberts (1974),
Bloomfield and Steiger (1980), Dodge (1997), Koenker (1997), Koenker and
d’Orey (1987), Portnoy (1997), and Portnoy and Koenker (1997). See Harter
(1974a,b, 1975a,b,c, 1976) for a historical account of linear regression. Draper
(2000) provides a bibliography of more recent references.
Early papers on transformations include Bartlett (1947) and Tukey (1957).
In a classic paper, Box and Cox (1964) developed numerical methods for es-
timating λo in the family of power transformations. It is well known that the
Box–Cox normal likelihood method for estimating λo can be sensitive to re-
mote or outlying observations. Cook and Wang (1983) suggested diagnostics
for detecting cases that influence the estimator, as did Tsai and Wu (1992),
Atkinson (1986), and Hinkley and Wang (1988). Yeo and Johnson (2000)
provide a family of transformations that does not require the variables to be
positive.
According to Tierney (1990, p. 297), one of the earliest uses of dynamic
graphics was to examine the effect of power transformations. In particular,
a method suggested by Fowlkes (1969) varies λ until the normal probability
plot is straight. McCulloch (1993) also gave a graphical method for finding
T
response transformations. A similar method would plot Y (λ) vs (α̂0 + β̂ λ x)
for λ ∈ Λ. See Example 1.5. Cook and Weisberg (1982, section 2.4) surveys
several transformation methods, and Cook and Weisberg (1994) described
how to use an inverse response plot of fitted values versus Y to visualize the
needed transformation.
The literature on numerical methods for variable selection in the OLS
multiple linear regression model is enormous. Three important papers are
Jones (1946), Mallows (1973), and Furnival and Wilson (1974). Chatterjee
and Hadi (1988, p. 43-47) give a nice account on the effects of overfitting
on the least squares estimates. Also see Claeskins and Hjort (2003), Hjort
and Claeskins (2003) and Efron, Hastie, Johnstone and Tibshirani (2004).
Some useful ideas for variable selection when outliers are present are given
by Burman and Nolan (1995), Ronchetti and Staudte (1994), and Sommer
and Huggins (1996).
In the variable selection problem, the FF and RR plots can be highly
CHAPTER 5. MULTIPLE LINEAR REGRESSION 167
5.5 Problems
Problems with an asterisk * are especially important.
5.1. Suppose that the regression model is Yi = 7+ βXi + ei for i = 1, ..., n
where the ei are iid N(0, σ 2 ) random variables. The least squares criterion
n
is Q(η) = (Yi − 7 − ηXi )2 .
i=1
a) What is E(Yi )?
b) Find the least squares estimator b of β by setting the first derivative
d
Q(η) equal to zero.
dη
c) Show that your b is the global minimizer of the least squares criterion
d2
Q by showing that the second derivative 2 Q(η) > 0 for all values of η.
dη
5.2. The location model is Yi = µ + ei for i = 1, ..., n where the ei are iid
with mean E(ei ) = 0 and constant variance VAR(ei ) = σ 2 . The least squares
n
estimator µ̂ of µ minimizes the least squares criterion Q(η) = (Yi − η)2.
i=1
CHAPTER 5. MULTIPLE LINEAR REGRESSION 168
b) Find E(b).
c) Find VAR(b).
n
(Hint: Note that b = i=1 ki Yi where the ki depend on the Xi which are
treated as constants.)
5.4. Assume that the response variable Y is height, and the explanatory
variables are X2 = sternal height, X3 = cephalic index, X4 = finger to ground,
X5 = head length, X6 = nasal height, X7 = bigonal breadth. Suppose that
the full model uses all 6 predictors plus a constant (= X1 ) while the reduced
model uses the constant and sternal height. Test whether the reduced model
can be used instead of the full model using the output on the previous page.
The data set had 74 cases.
5.5. The above output comes from the Johnson (1996) STATLIB data
set bodyfat after several outliers are deleted. It is believed that Y = β1 +
β2X2 + β3X22 + e where Y is the person’s bodyfat and X2 is the person’s
density. Measurements on 245 people were taken and are represented by
the output above. In addition to X2 and X22 , 7 additional measurements
X4 , ..., X10 were taken. Both the full and reduced models contain a constant
X1 ≡ 1.
a) Predict Y if X2 = 1.04. (Use the reduced model Y = β1 + β2X2 +
β3X22 + e.)
b) Test whether the reduced model can be used instead of the full model.
CHAPTER 5. MULTIPLE LINEAR REGRESSION 170
5.9. The output above was produced from the file mussels.lsp in Arc.
Let Y = log(M) where M is the muscle mass of a mussel. Let X1 ≡ 1, X2 =
log(H) where H is the height of the shell, and let X3 = log(S) where S is
the shell mass. Suppose that it is desired to predict Yh,new if log(H) = 4 and
log(S) = 5, so that xh = (1, 4, 5). Assume that se(Ŷh ) = 0.410715 and that
se(pred) = 0.467664.
a) If xh = (1, 4, 5) find a 99% confidence interval for E(Yh ).
b) If xh = (1, 4, 5) find a 99% prediction interval for Yh,new .
CHAPTER 5. MULTIPLE LINEAR REGRESSION 171
5.11. The output above is from the multiple linear regression of the
response y = height on the two nontrivial predictors sternal height = height
at shoulder and finger to ground = distance from the tip of a person’s middle
finger to the ground.
a) Consider the plot with yi on the vertical axis and the least squares
fitted values ŷi on the horizontal axis. Sketch how this plot should look if
the multiple linear regression model is appropriate.
b) Sketch how the residual plot should look if the residuals ri are on the
vertical axis and the fitted values ŷi are on the horizontal axis.
c) From the output, are sternal height and finger to ground useful for
predicting height? (Perform the ANOVA F test.)
5.12. Suppose that it is desired to predict the weight of the brain (in
grams) from the cephalic index measurement. The output below uses data
from 267 people.
5.17. The above table gives summary statistics for 4 MLR models con-
sidered as final submodels after performing variable selection. The forward
response plot and residual plot for the full model L1 was good. Model L3
was the minimum Cp model found. Which model should be used as the final
submodel? Explain briefly why each of the other 3 submodels should not be
used.
5.18. The above table gives summary statistics for 4 MLR models con-
sidered as final submodels after performing variable selection. The forward
response plot and residual plot for the full model L1 was good. Model L2
was the minimum Cp model found. Which model should be used as the final
submodel? Explain briefly why each of the other 3 submodels should not be
used.
CHAPTER 5. MULTIPLE LINEAR REGRESSION 174
5.19. The output above is from software that does all subsets variable
selection. The data is from Ashworth (1842). The predictors were A =
log(1692 property value), B = log(1841 property value) and C = log(percent
increase in value) while the response variable is Y = log(1841 population).
a) The top output corresponds to data with 2 small outliers. From this
output, what is the best model? Explain briefly.
b) The bottom output corresponds to the data with the 2 outliers re-
moved. From this output, what is the best model? Explain briefly.
CHAPTER 5. MULTIPLE LINEAR REGRESSION 175
nx <- matrix(rnorm(300),nrow=100,ncol=3)
ffL(nx,y)
Include the F F λ plot in Word by pressing the Ctrl and c keys simulta-
neously. This will copy the graph. Then in Word use the menu commands
“File>Paste”.
d) To make the transformation plots type the following command.
Tplt(nx,y)
The first plot will be for λ = −1. Move the curser to the plot and hold
the rightmost mouse key down. Highlight stop to go to the next plot.
Repeat these mouse operations to look at all of the plots. When you get a
plot that clusters about the OLS line which is included in each plot, include
CHAPTER 5. MULTIPLE LINEAR REGRESSION 176
this transformation plot in Word by pressing the Ctrl and c keys simulta-
neously. This will copy the graph. Then in Word use the menu commands
“File>Paste”. You should get the log transformation.
e) Type the following commands.
Use the mouse to highlight the created output and include the output in
Word.
) using the output in
f) Write down the least squares equation for log(Y
e).
Problems using ARC
To quit Arc, move curser to the x in the northeast corner and click.
Problems 5.21–5.26 use data sets that come with Arc (Cook and Weisberg
1999a).
5.21∗. a) In Arc enter the menu commands “File>Load>Data>ARCG”
and open the file big-mac.lsp. Next use the menu commands “Graph&Fit>
Plot of” to obtain a dialog window. Double click on TeachSal and then
double click on BigMac. Then click on OK. These commands make a plot
of x = TeachSal = primary teacher salary in thousands of dollars versus y =
BigMac = minutes of labor needed to buy a Big Mac and fries. Include the
plot in Word.
Consider transforming y with a (modified) power transformation
(y λ − 1)/λ, λ = 0
y (λ) =
log(y), λ=0
The response variable Y is the mussel muscle mass M, and the explanatory
variables are X2 = S = shell mass, X3 = H = shell height, X4 = L = shell
length and X5 = W = shell width.
Enter the menu commands “Graph&Fit>Fit linear LS” and fit the model:
enter S, H, L, W in the “Terms/Predictors” box, M in the “Response” box
and click on OK.
a) To get a forward response plot, enter the menu commands
“Graph&Fit>Plot of” and place L1:Fit-Values in the H–box and M in the
V–box. Copy the plot into Word.
b) Based on the forward response plot, does a linear model seem reason-
able?
c) To get a residual plot, enter the menu commands “Graph&Fit>Plot
of” and place L1:Fit-Values in the H–box and L1:Residuals in the V–box.
Copy the plot into Word.
d) Based on the residual plot, what MLR assumption seems to be vio-
lated?
e) Include the regression output in Word.
f) Ignoring the fact that an important MLR assumption seems to have
been violated, do any of predictors seem to be needed given that the other
predictors are in the model?
g) Ignoring the fact that an important MLR assumption seems to have
been violated, perform the ANOVA F test.
5.23∗. In Arc enter the menu commands “File>Load>Data>ARCG”
and open the file mussels.lsp. Use the commands “Graph&Fit>Scatterplot
Matrix of.” In the dialog window select H, L, W, S and M (so select M last).
Click on “OK” and include the scatterplot matrix in Word. The response M
is the edible part of the mussel while the 4 predictors are shell measurements.
Are any of the marginal predictor relationships nonlinear? Is E(M|H) linear
or nonlinear?
5.24∗. The file wool.lsp has data from a 33 experiment on the behavior of
worsted yarn under cycles of repeated loadings. The response y is the number
of cycles to failure and the three predictors are the length, amplitude and
CHAPTER 5. MULTIPLE LINEAR REGRESSION 178
d) Similarly make an FF plot using the fitted values from the two models.
Add the two lines. Include the plot in Word.
e) Next put the residuals from the submodel on the V axis and log(Ht)
on the H axis. Include this residual plot in Word.
f) Next put the residuals from the submodel on the V axis and the fitted
values from the submodel on the H axis. Include this residual plot in Word.
g) Next put log(Vol) on the V axis and the fitted values from the submodel
on the H axis. Include this forward response plot in Word.
h) Does log(Ht) seem to be an important term? If the only goal is to
predict volume, will much information be lost if log(Ht) is omitted? Remark
on the information given by each of the 6 plots. (Some of the plots
will suggest that log(Ht) is needed while others will suggest that log(Ht) is
not needed.)
5.26∗. a) In this problem we want to build a MLR model to predict
Y = g(BigMac) for some power transformation g. In Arc enter the menu
commands “File>Load>Data>Arcg” and open the file big-mac.lsp. Make
a scatterplot matrix of the variate valued variables and include the plot in
Word.
b) The log rule makes sense for the BigMac data. From the scatterplot,
use the “Transformations” menu and select “Transform to logs”. Include the
resulting scatterplot in Word.
c) From the “Mac” menu, select “Transform”. Then select all 10 vari-
ables and click on the “Log transformations” button. Then click on “OK”.
From the “Graph&Fit” menu, select “Fit linear LS.” Use log[BigMac] as the
response and the other 9 “log variables” as the Terms. This model is the full
model. Include the output in Word.
d) Make a forward response plot (L1:Fit-Values in H and log(BigMac) in
V) and residual plot (L1:Fit-Values in H and L1:Residuals in V) and include
both plots in Word.
e) Using the “L1” menu, select “Examine submodels” and try forward
selection and backward elimination. Using the Cp ≤ 2k rule suggests that the
submodel using log[service], log[TeachSal] and log[TeachTax] may be good.
From the “Graph&Fit” menu, select “Fit linear LS”, fit the submodel and
CHAPTER 5. MULTIPLE LINEAR REGRESSION 180
5.29∗. The following data set has 5 babies that are “good leverage
points:” they look like outliers but should not be deleted because they follow
the same model as the bulk of the data.
a) In Arc enter the menu commands “File>Load>3 1/2 Floppy(A:)” and
open the file cbrain.lsp. Select transform from the cbrain menu, and add
size1/3 using the power transformation option (p = 1/3). From
Graph&Fit, select Fit linear LS. Let the response be brnweight and as terms
include everything but size and Obs. Hence your model will include size1/3.
This regression will add L1 to the menu bar. From this menu, select Examine
submodels. Choose forward selection. You should get models including k =
2 to 12 terms including the constant. Find the model with the smallest
Cp (I) = CI statistic and include all models with the same k as that model
in Word. That is, if k = 2 produced the smallest CI , then put the block
with k = 2 into Word. Next go to the L1 menu, choose Examine submodels
and choose Backward Elimination. Find the model with the smallest CI and
include all of the models with the same value of k in Word.
b) What model was chosen by forward selection?
c) What model was chosen by backward elimination?
d) Which model do you prefer?
e) Give an explanation for why the two models are different.
f) Pick a submodel and include the regression output in Word.
g) For your submodel in f), make an RR plot with the residuals from the
CHAPTER 5. MULTIPLE LINEAR REGRESSION 182
full model on the V axis and the residuals from the submodel on the H axis.
Add the OLS line and the identity line y=x as visual aids. Include the RR
plot in Word.
h) Similarly make an FF plot using the fitted values from the two models.
Add the two lines. Include the FF plot in Word.
i) Using the submodel, include the forward response plot (of Ŷ versus Y )
and residual plot (of Ŷ versus the residuals) in Word.
j) Using results from f)-i), explain why your submodel is a good model.
5.30. a) In Arc enter the menu commands “File>Load>3 1/2 Floppy(A:)”
and open the file cyp.lsp. This data set consists of various measurements
taken on men from Cyprus around 1920. Let the response Y = height and
X = cephalic index = 100(head breadth)/(head length). Use Arc to get the
least squares output and include the relevant output in Word.
b) Intuitively, the cephalic index should not be a good predictor for a
person’s height. Perform a 4 step test of hypotheses with Ho: β2 = 0.
5.31. a) In Arc enter the menu commands “File>Load>3 1/2 Floppy(A:)”
and open the file cyp.lsp.
The response variable Y is height, and the explanatory variables are a
constant, X2 = sternal height (probably height at shoulder) and X3 = finger
to ground.
Enter the menu commands “Graph&Fit>Fit linear LS” and fit the model:
enter sternal height and finger to ground in the “Terms/Predictors” box,
height in the “Response” box and click on OK.
Include the output in Word. Your output should certainly include the
lines from “Response = height” to the ANOVA table.
b) Predict Y if X2 = 1400 and X3 = 650.
c) Perform a 4 step ANOVA F test of the hypotheses with
Ho: β2 = β3 = 0.
d) Find a 99% CI for β2.
e) Find a 99% CI for β3.
f) Perform a 4 step test for β2 = 0.
CHAPTER 5. MULTIPLE LINEAR REGRESSION 183
Regression Diagnostics
185
CHAPTER 6. REGRESSION DIAGNOSTICS 186
jth predictor Xj = (x1,j , ..., xn,j )T where j = k + 1. The sample mean and
covariance matrix of the nontrivial predictors are
1
n
u= ui (6.1)
n i=1
and
1
n
C = Cov(U ) = (ui − u)(ui − u)T , (6.2)
n − 1 i=1
respectively.
Some important numerical quantities that are used as diagnostics measure
≡β
the distance of ui from u and the influence of case i on the OLS fit β OLS .
Recall that the vector of fitted values =
= X(X T X)−1 X T Y = HY
Y = X β
where H is the hat matrix. Recall that the ith residual ri = Yi − Yi . Case (or
leave one out or deletion) diagnostics are computed by omitting the ith case
from the OLS regression. Following Cook and Weisberg (1999a, p. 357), let
(i)
Y (i) = X β (6.3)
denote the n × 1 vector of fitted values for estimating β with OLS without
the ith case. Denote the jth element of Y (i) by Y(i),j . It can be shown that
the variance of the ith residual VAR(ri ) = σ 2(1 − hi ). The usual estimator
of the error variance is n 2
2 r
= i=1 i .
σ
n−p
The (internally) studentized residual
ri
ei = √
1 − hi
σ
has zero mean and unit variance.
Definition 6.1. The ith leverage hi = H ii is the ith diagonal element of
the hat matrix H. The ith squared (classical) Mahalanobis distance
(β T X T X(β
(i) − β) (i) − β)
(Y (i) − Y )T (Y (i) − Y )
CDi = = (6.4)
σ2
p σ2
p
1
n
= 2 (Y(i),j − Yj )2 .
p
σ j=1
ri2 hi e2i hi
CDi = 2 = .
σ (1 − hi ) 1 − hi
p p 1 − hi
When the statistics CDi , hi and MDi are large, case i may be an outlier or
influential case. Examining a stem plot or dot plot of these three statistics for
unusually large values can be useful for flagging influential cases. Cook and
Weisberg (1999a, p. 358) suggest examining cases with CDi > 0.5 and that
cases with CDi > 1 should always be studied. Since H = H T and H = HH,
the hat matrix is symmetric and idempotent. Hence the eigenvalues of H
n
are zero or one and trace(H) = i=1 hi = p. Rousseeuw and Leroy (1987, p.
220 and p. 224) suggest using hi > 2p/n and MD2i > χ2p−1,0.95 as benchmarks
for leverages and Mahalanobis distances where χ2p−1,0.95 is the 95th percentile
of a chi–square distribution with p − 1 degrees of freedom.
Note that Proposition 6.1c) implies that Cook’s distance is the product
of the squared residual and a quantity that becomes larger the farther ui is
from u. Hence influence is roughly the product of leverage and distance of
Yi from Yi (see Fox 1991, p. 21). Mahalanobis distances and leverages both
define ellipsoids based on a metric closely related to the sample covariance
matrix of the nontrivial predictors. All points ui on the same ellipsoidal
CHAPTER 6. REGRESSION DIAGNOSTICS 188
contour are the same distance from u and have the same leverage (or the
same Mahalanobis distance).
Cook’s distances, leverages, and Mahalanobis distances can be effective
for finding influential cases when there is a single outlier, but can fail if there
are two or more outliers. Nevertheless, these numerical diagnostics combined
with plots such as residuals versus fitted values and fitted values versus the
response are probably the most effective techniques for detecting cases that
effect the fitted values when the multiple linear regression model is a good
approximation for the bulk of the data. In fact, these diagnostics may be
useful for perhaps up to 90% of such data sets while residuals from robust
regression and Mahalanobis distances from robust estimators of multivariate
location and dispersion may be helpful for perhaps another 3% of such data
sets.
263-266) of age less than 7 months that are x-outliers. Nine toddlers were
between 7 months and 3.5 years of age, four of whom appear to be x-outliers
(cases 241, 243, 267, and 269). (The points are not labeled on the plot, but
the five infants and these four toddlers are easy to recognize when discrepant.)
Figure 1.1 (on p. 7) shows the RR plot. We dispose of the OLS and
L1 fits by noting that the very close agreement in their residuals implies
an operational equivalence in the two fits. ALMS fits the nine x-outliers
quite differently than OLS, L1 , and ALTS. All fits are highly correlated for
the remaining 265 points, showing that all fits agree on these cases, thus
focusing attention on the infants and toddlers.
All of the Splus fits except ALMS accommodated the infants. The funda-
mental reason that ALMS is the “outlier” among the fits is that the infants
and toddlers, while well separated from the rest of data, turn out to fit the
overall linear model quite well. A strength of the LMS criterion – that it
does not pay much attention to the leverage of cases – is perhaps a weakness
here since it leads to the impression that these cases are bad, whereas they
are no more than atypical.
Turning to optimization issues, ALMS had an objective function of 52.7
while KLMS had a much higher objective function of 114.7 even though
KLMS used ten times as many subsamples. The current version of lmsreg
will no longer give a smaller objective function to the algorithm that uses a
smaller number of subsamples.
Figure 1.2 (on p. 8) shows the residual plots for the Gladstone data when
one observation, 119, had head length entered incorrectly as 109 instead of
199. Unfortunately, the ALMS estimator did not detect this outlier.
Example 6.2. Buxton (1920, p. 232-5) gives 20 measurements of 88
men. We chose to predict stature using an intercept, head length, nasal
height, bigonal breadth, and cephalic index. Observation 9 was deleted since
it had missing values. Five individuals, numbers 62-66, were reported to be
about 0.75 inches tall with head lengths well over five feet! This appears to
be a clerical error; these individuals’ stature was recorded as head length and
the integer 18 or 19 given for stature, making the cases massive outliers with
enormous leverage. These absurdly bad observations turned out to confound
the standard high breakdown (HB) estimators. The residual plots in Figure
1.3 (on p. 10) show that five of the six Splus estimators accommodated them.
This is a warning that even using the objective of high breakdown will not
necessarily protect one from extremely aberrant data. Nor should we take
CHAPTER 6. REGRESSION DIAGNOSTICS 192
much comfort in the fact that KLMS clearly identified them; the criterion of
this fit was worse than that of the ALMS fit, and so should be regarded as
inferior.
This plot is no longer reproducible because of changes in the Splus code.
Figure 7.1 (on p. 229) shows the RR plot for more current (as of 2000) Splus
implementations of lmsreg and ltsreg. Problem 6.1 shows how to create
RR and FF plots.
Example 6.3. Figure 1.6 (on p. 17) is nearly identical to a forward
response plot. Since the plotted points do not scatter about the identity
line, the multiple linear regression model is not appropriate. Nevertheless,
Yi ∝ (xTi β̂ OLS )3 .
In Chapter 12 it will be shown that the forward response plot is useful for
visualizing the conditional distribution Y |βT x in 1D regression models where
Y x|β T x.
3
1800
1700
Y
63
1600
44
1500
FIT
Residual Plot
3
50
0
RES
63
-50
-100
44
FIT
Figure 6.1: Residual and Forward Response Plots for the Tremearne Data
CHAPTER 6. REGRESSION DIAGNOSTICS 194
The following techniques are useful for detecting outliers when the mul-
tiple linear regression model is appropriate.
1. Find the OLS residuals and fitted values and make a forward response
plot and a residual plot. Look for clusters of points that are separated
from the bulk of the data and look for residuals that have large absolute
values. Beginners frequently label too many points as outliers. Try
to estimate the standard deviation of the residuals in both plots. In
the residual plot, look for residuals that are more than 5 standard
deviations away from the r = 0 line.
2. Make an RR plot. See Figures 1.1 and 7.1 on p. 7 and p. 229, respec-
tively.
4. Display the residual plots from several different estimators. See Figures
1.2 and 1.3 on p. 8 and p. 10, respectively.
y = m(βT x) + e. (6.7)
and thus
yi = t(zi ) = β T xi + ei . (6.9)
There are several important regression models that do not have additive
errors including generalized linear models. If
y = g(βT x, e) (6.10)
T
β̂ x versus y if model (6.10) holds. Residual plots are also used for model
assessment, but residual plots emphasize lack of fit.
The following notation is useful. Let m̂ be an estimator of m. Let the
ith predicted or fitted value ŷi = m̂i = m̂(xTi ), and let the ith residual
ri = yi − ŷi .
Definition 6.6. Then a fit–response or FY plot is a plot of ŷ versus y.
Application 6.1. Use the FY plot to check the model for goodness of
fit, outliers and influential cases.
To understand the information contained in the FY plot, first consider a
plot of mi versus yi . Ignoring the error in the model yi = mi + ei gives y = m
which is the equation of the identity line with unit slope and zero intercept.
The vertical deviations from the identity line are yi − mi = ei . The reasoning
for the FY plot is very similar. The line y = ŷ is the identity line and the
vertical deviations from the line are the residuals yi − m̂i = yi − ŷi = ri .
Suppose that the model yi = mi + ei is a good approximation to the data
and that m̂ is a good estimator of m. If the identity line is added to the plot
as a visual aid, then the plotted points will scatter about the line and the
variability of the residuals can be examined.
For a given data set, it will often be useful to generate the FY plot,
residual plots, and model checking plots. An advantage of the FY plot is
that if the model is not a good approximation to the data or if the estimator
m̂ is poor, then detecting deviations from the identity line is simple. Also,
residual variability is easier to judge against a line than a curve. On the
other hand, model checking plots may provide information about the form
of the conditional mean function E(y|x) = m(xT ). See Chapter 12.
Many numerical diagnostics for detecting outliers and influential cases on
the fit have been suggested, and often this research generalizes results from
Cook (1977, 1986) to various models of form (6.6). Information from these
diagnostics can be incorporated into the FY plot by highlighting cases that
have large absolute values of the diagnostic.
The most important example is the multiple linear regression (MLR)
model. For this model, the FY plot is the forward response plot. If the MLR
model holds and the errors ei are iid with zero mean and constant variance
σ 2, then the plotted points should scatter about the identity line with no
other pattern.
When the bulk of the data follows the MLR model, the following rules
CHAPTER 6. REGRESSION DIAGNOSTICS 198
of thumb are useful for finding influential cases and outliers. Look for points
with large absolute residuals and for points far away from y. Also look for
gaps separating the data into clusters. To determine whether small clusters
are outliers or good leverage points, give zero weight to the clusters, and fit
a MLR estimator to the bulk of the data. Denote the weighted estimator by
β̂w . Then plot ŷw versus y using the entire data set. If the identity line passes
through the bulk of the data but not the cluster, then the cluster points may
be outliers.
To see why gaps are important, suppose that OLS was used to obtain
ŷ = m̂. Then the squared correlation (corr(y, ŷ))2 is equal to the coefficient
of determination R2 . Even if an alternative MLR estimator is used, R2 over
emphasizes the strength of the MLR relationship when there are two clusters
of data since much of the variability of y is due to the smaller cluster.
Now suppose that the MLR model is incorrect. If OLS is used in the FY
plot, and if y = g(βT x, e), then the plot can be used to visualize g for many
data sets (see Ch. 12). Hence the plotted points may be very far from linear.
The plotted points in FY plots created from other MLR estimators may not
be useful for visualizing g, but will also often be far from linear.
A commonly used diagnostic is Cook’s distance CDi . Assume that OLS
is used to fit the model and to make the FY plot ŷ versus y. Then CDi
tends to be large if ŷ is far from the sample mean y and if the corresponding
absolute residual |ri | is not small. If ŷ is close to y then CDi tends to be
small unless |ri | is large. An exception to these rules of thumb occurs if a
group of cases form a cluster and the OLS fit passes through the cluster.
Then the CDi ’s corresponding to these cases tend to be small even if the
cluster is far from y.
An advantage of the FY plot over numerical diagnostics is that while it
depends strongly on the model m, defining diagnostics for different fitting
methods can be difficult while the FY plot is simply a plot of ŷ versus y. For
the MLR model, the FY plot can be made from any good MLR estimator,
including OLS, least absolute deviations and the R/Splus estimator lmsreg.
Example 6.2 (continued): Figure 6.2 shows the forward response plot
and residual plot for the Buxton data. Although an index plot of Cook’s
distance CDi may be useful for flagging influential cases, the index plot
provides no direct way of judging the model against the data. As a remedy,
cases in the FY plot with CDi > min(0.5, 2p/n) were highlighted. Notice
CHAPTER 6. REGRESSION DIAGNOSTICS 199
1500
1000
Y
500
63 65
61 64
62
0
FIT
Residual Plot
100
61
50
64
RES
63
65
-50
62
-150
FIT
that the OLS fit passes through the outliers, but the FY plot is resistant to y–
outliers since y is on the vertical axis. Also notice that although the outlying
cluster is far from y only two of the outliers had large Cook’s distance. Hence
masking occurred for both Cook’s distances and for OLS residuals, but not
for OLS fitted values. FY plots using other MLR estimators such as lmsreg
were similar.
High leverage outliers are a particular challenge to conventional numer-
ical MLR diagnostics such as Cook’s distance, but can often be visualized
using the forward response and residual plots. (Using the trimmed views of
Section 11.3 and Chapter 12 is also effective for detecting outliers and other
departures from the MLR model.)
Example 6.5. Hawkins, Bradu, and Kass (1984) present a well known
artificial data set where the first 10 cases are outliers while cases 11-14 are
good leverage points. Figure 6.3 shows the residual and forward response
plots based on the OLS estimator. The highlighted cases have Cook’s dis-
tance > min(0.5, 2p/n), and the identity line is shown in the FY plot. Since
the good cases 11-14 have the largest Cook’s distances and absolute OLS
CHAPTER 6. REGRESSION DIAGNOSTICS 200
7 8
10 5
10
2 6 3
1 9 4
8
6
Y
13
2
14 11
12
0
0 2 4 6 8
FIT
Residual Plot
7 8
4
2 6 10 5
1 9 3
2
4
0
RES
-2
-4
14
-6
13
12
-8
11
0 2 4 6 8
FIT
FY PLOT
1.0
0.5
0.0
Y
-0.5
-1.0
5 2
4 3
-1.5
10 8
9 7
1
6
-2.0
FIT
9
8
7
Y
6
5
4
4 5 6 7 8 9
FIT
residuals, swamping has occurred. (Masking has also occurred since the out-
liers have small Cook’s distances, and some of the outliers have smaller OLS
residuals than clean cases.) To determine whether both clusters are outliers
or if one cluster consists of good leverage points, cases in both clusters could
be given weight zero and the resulting forward response plot created. (Al-
ternatively, forward response plots based on the tvreg estimator of Section
11.3 could be made with the untrimmed cases highlighted. For high levels of
trimming, the identity line often passes through the good leverage points.)
The above example is typical of many “benchmark” outlier data sets for
MLR. In these data sets traditional OLS diagnostics such as Cook’s distance
and the residuals often fail to detect the outliers, but the combination of the
FY plot and residual plot is usually able to detect the outliers.
Example 6.6. MathSoft (1999a, p. 245-246) gives an FY plot for sim-
ulated data. In this example the simulated data set is modified by planting
10 outliers. Let x1 and x2 be iid uniform U(−1, 1) random variables, and let
y = x1 x2 + e where the errors e are iid N(0, 0.04) random variables. The ar-
tificial data set uses 400 cases, but the first 10 cases used y ∼ N(−1.5, 0.04),
CHAPTER 6. REGRESSION DIAGNOSTICS 202
x1 ∼ N(0.2, 0.04) and x2 ∼ N(0.2, 0.04) where y, x1, and x2 were indepen-
dent. The model y = m(x1, x2 ) + e was fitted nonparametrically without
using knowledge of the true regression relationship. The fitted values m̂
were obtained from the Splus function ppreg for projection pursuit regres-
sion (Friedman and Stuetzle, 1981). The outliers are easily detected with
the FY plot shown in Figure 6.4.
Example 6.7. The lynx data is a well known time series concerning the
number wi of lynx trapped in a section of Northwest Canada from 1821 to
1934. There were n = 114 cases and MathSoft (1999b, p. 166-169) fits an
AR(11) model yi = β0 + β1yi−1 + β2yi−2 + · · · + β11yi−11 + ei to the data
where yi = log(wi ) and i = 12, 13, ..., 114. The FY plot shown in Figure 6.5
suggests that the AR(11) model fits the data reasonably well. To compare
different models or to find a better model, use an FF plot of Y and the fitted
values from several competing time series models. See Problem 6.4.
6.5 Complements
Excellent introductions to OLS diagnostics include Fox (1991) and Cook and
Weisberg (1999a, p. 161-163, 183-184, section 10.5, section 10.6, ch. 14, ch.
15, ch. 17, ch. 18, and section 19.3). More advanced works include Belsley,
Kuh, and Welsch (1980), Cook and Weisberg (1982), Atkinson (1985) and
Chatterjee and Hadi (1988). Hoaglin and Welsh (1978) examines the hat
matrix while Cook (1977) introduces Cook’s distance.
Some other papers of interest include Barrett and Gray (1992), Gray
(1985), Hadi and Simonoff (1993), Hettmansperger and Sheather (1992),
Velilla (1998), and Velleman and Welsch (1981).
Hawkins and Olive (2002, p. 141, 158) suggest using the RR and FF
plots. Typically RR and FF plots are used if there are several estimators for
one fixed model, eg OLS versus L1 or frequentist versus Bayesian for multiple
linear regression, or if there are several competing models. An advantage of
the FF plot is that the response Y can be added to the plot. The FFλ
plot is an FF plot where the fitted values were obtained from competing
power transformation models indexed by the power transformation parameter
λ ∈ Λc . Variable selection uses both FF and RR plots.
Rousseeuw and van Zomeren (1990) suggest that Mahalanobis distances
CHAPTER 6. REGRESSION DIAGNOSTICS 203
6.6 Problems
R/Splus Problems
Warning: Use the command source(“A:/rpack.txt”) to download
the programs and the command source(“A:/robdata.txt”) to download
the data. See Preface or Section 14.2. Typing the name of the rpack
function, eg MLRplot, will display the code for the function. Use the args
command, eg args(MLRplot), to display the needed arguments for the func-
tion.
6.1∗. a) After entering the two source commands above, enter the follow-
ing command.
> MLRplot(buxx,buxy)
CHAPTER 6. REGRESSION DIAGNOSTICS 204
Click the rightmost mouse button (and in R click on Stop). The forward
response plot should appear. Again, click the rightmost mouse button (and
in R click on Stop). The residual plot should appear. Hold down the Ctrl
and c keys to make a copy of the two plots. Then paste the plots in Word.
b) The response variable in height, but 5 cases were recorded with heights
about 0.75 inches tall. The highlighted squares in the two plots correspond
to cases with large Cook’s distances. With respect to the Cook’s distances,
what is happening, swamping or masking?
c) RR plots: One feature of the MBA estimator (see Chapter 7) is that it
depends on the sample of 7 centers drawn and changes each time the function
is called. In ten runs, about nine plots will look like Figure 7.1, but in about
one plot the MBA estimator will also pass through the outliers.
If you use R, type the following command and include the plot in Word.
> library(lqs)
> rrplot2(buxx,buxy)
If you use Splus, type the following command and include the plot in
Word.
> rrplot(buxx,buxy)
d) FF plots: ideally, the plots in the top row will cluster about the identity
line.
If you use R, type the following command and include the plot in Word.
> library(lqs)
> ffplot2(buxx,buxy)
If you use Splus, type the following command and include the plot in
Word.
> ffplot(buxx,buxy)
> diagplot(buxx,buxy)
6.3. This problem makes the FY plot for the lynx data in Example 6.7.
a) Check that the lynx data is in Splus by typing the command help(lynx).
A window will appear if the data is available.
b) For Splus, enter the following Splus commands to produce the FY plot.
Include the plot in Word. The command abline(0,1) adds the identity line.
> library(ts)
> data(lynx)
> Y <- log(lynx)
> out <- ar.yw(Y)
> Yts <- Y[12:114]
> FIT <- Yts - out$resid[12:114]
> plot(FIT,Yts)
> abline(0,1)
6.4∗. Following Lin and Pourahmadi (1998), consider the lynx time se-
ries data and let the response Yt = log(lynx). Moran (1953) suggested the
autoregressive AR(2) model Ŷt = 1.05 + 1.41Yt−1 − 0.77Yt−2 . Tong (1977)
suggested the AR(11) model Ŷt = 1.13Yt−1 − .51Yt−2 + .23Yt−3 − 0.29Yt−4 +
.14Yt−5 − 0.14Yt−6 + .08Yt−7 − .04Yt−8 + .13Yt−9 + 0.19Yt−10 − .31Yt−11 . Brock-
well and Davis (1991, p. 550) suggested the AR(12) model Ŷt = 1.123 +
1.084Yt−1 − .477Yt−2 + .265Yt−3 − 0.218Yt−4 + .180Yt−9 − 0.224Yt−12 . Tong
(1983) suggested the following two self–exciting autoregressive models. The
SETAR(2,7,2) model uses Ŷt = .546 + 1.032Yt−1 − .173Yt−2 + .171Yt−3 −
0.431Yt−4 + .332Yt−5 − 0.284Yt−6 + .210Yt−7 if Yt−2 ≤ 3.116 and Ŷt = 2.632 +
1.492Yt−1 − 1.324Yt−2 , otherwise. The SETAR(2,5,2) model uses Ŷt = .768 +
1.064Yt−1 − .200Yt−2 + .164Yt−3 − 0.428Yt−4 + .181Yt−5 if Yt−2 ≤ 3.05 and
CHAPTER 6. REGRESSION DIAGNOSTICS 206
Ŷt = 2.54 + 1.474Yt−1 − 1.202Yt−2 , otherwise. The FF plot of the fitted val-
ues and the response can be used to compare the models. Type the rpack
command fflynx() (in R, 1st type library(ts) and data(lynx)).
a) Include the resulting plot and correlation matrix in Word.
b) Which model seems to be best? Explain briefly.
c) Which two pairs of models gave very similar fitted values?
6.5. This problem may not work in R. Type help(ppreg) to make
sure that Splus has the function ppreg. Then make the FY plot for Example
6.6 with the following commands. Include the plot in Word.
> set.seed(14)
> x1 <- runif(400,-1,1)
> x2 <- runif(400,-1,1)
> eps <- rnorm(400,0,.2)
> Y <- x1*x2 + eps
> x <- cbind(x1,x2)
> x[1:10,] <- rnorm(20,.2,.2)
> Y[1:10] <- rnorm(10,-1.5,.2)
> out <- ppreg(x,Y,2,3)
> FIT <- out$ypred
> plot(FIT,Y)
> abline(0,1)
Arc problems
Warning: The following problem uses data from the book’s web-
page. Save the data files on a disk. Get in Arc and use the menu com-
mands “File > Load” and a window with a Look in box will appear. Click
on the black triangle and then on 3 1/2 Floppy(A:). Then click twice on the
data set name.
Using material learned in Chapters 5–6, analyze the data sets described
in Problems 6.5–6.12. Assume that the response variable Y = t(Z) and
that the predictor variable X2 , ..., Xp are functions of remaining variables
W2 , ..., Wr. Unless told otherwise, the full model Y, X1 , X2 , ..., Xp (where
X1 ≡ 1) should use functions of every variable W2 , ..., Wr (and often p = r +
1). (In practice, often some of the variables and some of the cases are deleted,
but we will use all variables and cases, unless told otherwise, primarily so
CHAPTER 6. REGRESSION DIAGNOSTICS 207
that the instructor has some hope of grading the problems in a reasonable
amount of time.) See pages 163–165 for useful tips for building a full model.
Read the description of the data provided by Arc. Once you have a
good full model, perform forward selection and backward elimination. Find
the model that minimizes Cp (I) and find the smallest value of k such that
Cp (I) ≤ 2k. The minimum Cp model often has too many terms while the
2nd model sometimes has too few terms.
a) Give the output for your full model, including Y = t(Z) and R2 . If it
is not obvious from the output what your full model is, then write down the
full model. Include a forward response plot for the full model. (This plot
should be linear).
b) Give the output for your final submodel. If it is not obvious from the
output what your submodel is, then write down the final submodel.
c) Give between 3 and 5 plots that justify that your multiple linear re-
gression submodel is reasonable. Below or beside each plot, give a brief
explanation for how the plot gives support for your model.
6.6. For the file bodfat.lsp, described in Example 1.4, use Z = Y but do
not use X1 as a predictor in the full model. Do parts a), b) and c) above.
6.7∗. For the file boston2.lsp, described in Examples 1.6, 12.7 and 12.8
use Z = (y =)CRIM. Do parts a), b) and c) above Problem 6.6.
Note: Y = log(CRIM), X4 , X8 , is an interesting submodel, but more
predictors are probably needed.
6.8∗. For the file major.lsp, described in Example 6.4, use Z = Y . Do
parts a), b) and c) above Problem 6.6.
Note: there are 1 or more outliers that affect numerical methods of vari-
able selection.
6.9. For the file marry.lsp, described below, use Z = Y . This data set
comes from Hebbler (1847). The census takers were not always willing to
count a woman’s husband if he was not at home. Do not use the predictor
X2 in the full model. Do parts a), b) and c) above Problem 6.6.
6.10∗. For the file museum.lsp, described below, use Z = Y . Do parts
a), b) and c) above Problem 6.6.
This data set consists of measurements taken on skulls at a museum and
was extracted from tables in Schaaffhausen (1878). There are at least three
groups of data: humans, chimpanzees and gorillas. The OLS fit obtained
from the humans passes right through the chimpanzees. Since Arc numbers
CHAPTER 6. REGRESSION DIAGNOSTICS 208
cases starting at 0, cases 47–59 are apes. These cases can be deleted by
highlighting the cases with small values of Y in the scatterplot matrix and
using the case deletions menu. (You may need to maximize the window
containing the scatterplot matrix in order to see this menu.)
i) Try variable selection all of the data.
ii) Try variable selection without the apes.
√ perhaps only X1 , X2 and X3 should be used
If all of the cases are used,
in the full model. Note that Y and X2 have high correlation.
6.11∗. For the file pop.lsp, described below, use Z = Y . Do parts a), b)
and c) above Problem 6.6.
This data set comes from Ashworth (1842). Try transforming all variables
to logs. Then the added variable plots show two outliers. Delete these
two cases. Notice the effect of these two outliers on the p–values for the
coefficients and on numerical methods for variable selection.
Note: then log(Y ) and log(X2 ) make a good submodel.
6.12∗. For the file pov.lsp, described below, use i) Z = flife and ii)
Z = gnp2 = gnp + 2. This dataset comes from Rouncefield (1995). Making
loc into a factor may be a good idea. Use the commands poverty>Make
factors and select the variable loc. For ii), try transforming to logs and
deleting the 6 cases with gnp2 = 0. (These cases had missing values for gnp.
The file povc.lsp has these cases deleted.) Try your final submodel on the
data that includes the 6 cases with gnp2 = 0. Do parts a), b) and c) above
Problem 6.6.
6.13∗. For the file skeleton.lsp, described below, use Z = y. Do parts a),
b) and c) above Problem 6.6.
This data set is also from Schaaffhausen (1878). At one time I heard
or read a conversation between a criminal forensics expert with his date. It
went roughly like “If you wound up dead and I found your femur, I could tell
what your height was to within an inch.” Two things immediately occurred
to me. The first was “no way” and the second was that the man must not
get many dates! The files cyp.lsp and major.lsp have measurements including
height, but their R2 ≈ 0.9. The skeleton data set has at least four groups:
stillborn babies, newborns and children, older humans and apes.
a) Take logs of each variable and fit the regression on log(Y) on log(X1 ),
..., log(X13 ). Make a residual plot and highlight the case with the with the
smallest residual. From the Case deletions menu, select Delete selection from
CHAPTER 6. REGRESSION DIAGNOSTICS 209
file response Y
a) allomet.lsp BRAIN
b) casuarin.lsp W
c) evaporat.lsp Evap
d) hald.lsp Y
e) haystack.lsp Vol
CHAPTER 6. REGRESSION DIAGNOSTICS 210
f) highway.lsp rate
(from the menu Highway, select ‘‘Add a variate" and type
sigsp1 = sigs + 1. Then you can transform sigsp1.)
g) landrent.lsp Y
h) ozone.lsp ozone
i) paddle.lsp Weight
j) sniffer.lsp Y
k) water.lsp Y
i) Write down the full model that you use and include the full model
residual plot and forward response plot in Word. Give R2 for the full model.
Y = Xβ + e
is appropriate for all or for the bulk of the data. For a high breakdown (HB)
regression estimator b of β, the median absolute residual
stays bounded even if close to half of the data set cases are replaced by
arbitrarily bad outlying cases; ie, the breakdown value of the regression esti-
mator is close to 0.5. The concept of breakdown will be made more precise
in Section 9.4.
Perhaps the first HB regression estimator proposed was the least median
of squares (LMS) estimator. Let |r(b)|(i) denote the ith ordered absolute
2
residual from the estimate b sorted from smallest to largest, and let r(i) (b)
denote the ith ordered squared residual. Three of the most important robust
estimators are defined below.
Definition 7.1. The least quantile of squares (LQS(cn )) estimator mini-
mizes the criterion
2
QLQS (b) = r(c n)
(b). (7.1)
211
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 212
When cn /n → 1/2, the LQS(cn ) estimator is also known as the least median
of squares estimator (Hampel 1975, p. 380).
Definition 7.2. The least trimmed sum of squares (LTS(cn )) estimator
(Rousseeuw 1984) minimizes the criterion
cn
2
QLT S (b) = r(i) (b). (7.2)
i=1
These three estimators all find a set of fixed size cn = cn (p) ≥ n/2 cases
to cover, and then fit a classical estimator to the covered cases. LQS uses
the Chebyshev fit, LTA uses L1 , and LTS uses OLS.
Definition 7.4. The integer valued parameter cn is the coverage of the
estimator. The remaining n−cn cases are given weight zero. In the literature
and software,
cn = n/2
+ (p + 1)/2
(7.4)
is often used as the default. Here x
is the greatest integer less than or
equal to x. For example, 7.7
= 7.
Remark 7.1. Warning: In the literature, HB regression estimators
seem to come in two categories. The first category consists of estimators
that have no rigorous asymptotic theory but can be computed for very small
data sets. The second category consists of estimators that have rigorous
asymptotic theory but are impractical to compute. Due to the high compu-
tational complexity of these estimators, they are rarely used; however, the
criterion are widely used for fast approximate algorithm estimators that can
detect certain configurations of outliers. These approximations are typically
inconsistent estimators with low breakdown. One of the most disappointing
aspects of robust literature is that frequently no distinction is made between
the impractical HB estimators and the inconsistent algorithm estimators used
to detect outliers. Chapter 8 shows
√ how to fix some of these algorithms so
that the resulting estimator is n consistent and high breakdown.
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 213
where MED(n) is the sample median and MAD(n) is the sample median
absolute deviation. A convenient value for the trimming constant is k = 6.
Next, find the percentage of cases trimmed to the left and to the right by
Mn , and round both percentages up to the nearest integer and compute the
corresponding trimmed mean. Let TA,n denote the resulting estimator. For
example, if Mn trimmed the 7.3% smallest cases and the 9.76% largest cases,
then the final estimator TA,n is the (8%, 10%) trimmed mean. TA,n is asymp-
totically equivalent to a sequence of trimmed means with an asymptotic
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 214
n
n
2
Cn (b) = I[|r|(i)(b) ≤ k |r|(cn )(b)] = I[r(i) (b) ≤ k 2 r(c
2
n)
(b)]. (7.5)
i=1 i=1
The two stage trimmed mean inherits the breakdown value of the median
and the stability of a trimmed mean with a low trimming proportion. The
RLTx estimator can be regarded as an extension of the two stage mean to
regression. The RLTx estimator inherits the high breakdown value of the
ALTx(0.5) estimator, and the stability of the ALTx(τR ) where τR is typically
close to one.
The tuning parameter k ≥ 1 controls the amount of trimming. The in-
equality k ≥ 1 implies that Cn ≥ cn , so the RLTx(k) estimator generally has
higher coverage and therefore higher statistical efficiency than ALTx(0.5).
Notice that although L estimators ALTx(cn,j ) were defined, only two are
needed: ALTx(0.5) to get a resistant scale and define the coverage needed,
and the final estimator ALTX(τR ). The computational load is typically less
than twice that of computing the ALTx(0.5) estimator since the computa-
tional complexity of the ALTx(τ ) estimators decreases as τ increases from
0.5 to 1.
The behavior of the RLTx estimator is easy to understand. Compute
the most resistant ALTx estimator β̂ ALT x(cn ) and obtain the corresponding
residuals. Count the number Cn of absolute residuals that are no larger than
k |r|(cn ) ≈ kMED(|r|i ). Then find τR ∈ G and compute the RLTx estimator.
(The RLTx estimator uses Cn in a manner analogous to the way that the two
stage mean uses kMAD(n).) If k = 6, and the regression model holds, the
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 215
RLTx estimator will be the classical estimator or the ALTx estimator with
99% coverage for a wide variety of data sets. On the other hand, if β̂ ALT x(cn )
fits cn cases exactly, then |r|(cn ) = 0 and RLTx = ALTx(cn ).
The RLTx estimator has the same breakdown point as the ALTx(0.5)
estimator. Theoretical results and a simulation study, based on Olive and
Hawkins (2003) and presented in Sections 7.4 and 7.5, suggest that the RLTx
estimator is simultaneously more stable and more resistant than the ALTx(
0.75 n) estimator for x = A and S. Increasing the coverage for the LQS
criterion is not suggested since the Chebyshev fit tends to have less efficiency
than the LMS fit.
The least adaptively trimmed sum of squares (LATS(k)) estimator is the OLS
fit that minimizes
Cn (b)
2
QLAT S (b) = r(i) (b).
i=1
Note that the adaptive estimators reduce to the highest breakdown ver-
sions of the fixed coverage estimators if k = 1 and (provided there is no exact
fit to at least cn of the cases) to the classical estimators if k = ∞.
These three adaptive coverage estimators simultaneously achieve a high
breakdown value with high coverage, as do the RLTx estimators, but there
are important outlier configurations where the resistance of the two estima-
tors differs. The notation LATx will sometimes be used.
when
XT X
→ W −1 ,
n
and when the errors ei are iid with a cdf F and a unimodal pdf f that is
symmetric with a unique maximum at 0. When the variance V (ei ) exists,
1
V (OLS, F ) = V (ei) = σ 2 while V(L1 , F) = .
4[f(0)]2
See Koenker and Bassett (1978) and Bassett and Koenker (1978). Broffitt
(1974) compares OLS, L1 , and L∞ in the location model and shows that the
rate of convergence of the Chebyshev estimator is often very poor.
Remark 7.2. Obtaining asymptotic theory for LTA and LTS is a very
challenging problem. Mašı̈ček (2004) shows that LTS is consistent, but there
may be no other results outside of the location model. See Davies (1993),
Garcı́a-Escudero, Gordaliza and Matrán (1999), Hawkins and Olive (2002),
Hössjer (1994), Stromberg, Hawkins and Hössjer (2000), and Rousseeuw
(1984) for further discussion and conjectures. For the location model, Yohai
and Maronna (1976) and Butler (1982) derived asymptotic theory for LTS
while Tableman (1994ab) derived asymptotic theory for LTA. Shorack (1974)
and Shorack and Wellner (1986, section 19.3) derived the asymptotic theory
for a large class of location estimators that use random coverage (as do many
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 217
Cn (β̂ n ) P
→ τF ≡ τF (k) = F (k F −1(0.75)) − F (−k F −1 (0.75)). (7.7)
n
Proof. First assume that the predictors are bounded. Hence x ≤ M
for some constant M. Let 0 < γ < 1, and let 0 <
< 1. Since β̂n is
consistent, there exists an N such that
P (A) = P (β̂j,n ∈ [βj − , βj + ], j = 1, ..., p) ≥ 1 − γ
4pM 4pM
for all n ≥ N. If n ≥ N, then on set A,
p
sup |ri − ei| = sup | xi,j (βj − β̂j,n )| ≤ .
i=1,...,n i=1,...,n
i=1
2
1
n
Cn (β̂ n )
I[−kMED(|e1|) +
≤ ei ≤ kMED(|e1 |) −
] ≤
n i=1 n
1
n
≤ I[−kMED(|e1|) −
≤ ei ≤ kMED(|e1|) +
],
n i=1
and the result follows since γ and
are arbitrary and the three terms above
converge to τF almost surely as
goes to zero.
When x is bounded in probability, fix M and suppose Mn of the cases
have predictors xi such that xi ≤ M. By the argument above, the propor-
tion of absolute residuals of these cases that are below |r|(cMn ) converges in
probability to τF . But the proportion of such cases can be made arbitrarily
close to one as n increases to ∞ by increasing M. QED
Under the same conditions of Lemma 7.1,
β̂RLT x − β = OP (n−δ ).
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 219
Proof. Since G is finite, this result follows from Pratt (1959). QED
Theorem 7.3 shows that the RLTx estimator is asymptotically equivalent
to an ALTx estimator that typically has high coverage.
Theorem 7.3. Assume that τj , τj+1 ∈ G. If
P
P [Cn (β̂ ALT x(0.5))/n ∈ (τj , τj+1 )] → 1,
increase k. For example, similar statements hold for distributions with lighter
tails than the double exponential distribution if k ≥ 10.0 and n < 200.
Proposition 7.5: Breakdown of LTx, RLTx, and LATx Estima-
tors. LMS(τ ), LTS(τ ), and LTA(τ ) have breakdown value
min(1 − τ, τ ).
The breakdown value for the LATx estimators is 0.5, and the breakdown
value for the RLTx estimators is equal to the breakdown value of the ALTx(cn )
estimator.
The breakdown results for the LTx estimators are well known. See Hössjer
(1994, p. 151). Breakdown proofs in Rousseeuw and Bassett (1991) and
Niinimaa, Oja, and Tableman (1990) could also be modified to give the result.
See Section 9.4 for the definition of breakdown.
Theorem 7.6. The LMS(τ ) converges at a cubed root rate to a non-
Gaussian limit (under regularity conditions on the error distribution that
exclude the uniform distribution).
The proof of Theorem 7.6 is given in Davies (1990) and Kim and Pollard
(1990). Also see Davies (1993, p. 1897).
Conjecture 7.1. Let the iid errors ei have a cdf F that is continuous
and strictly increasing on its interval support with a symmetric, unimodal,
differentiable density f that strictly decreases as |x| increases on the support.
a) The estimator β̂ LT S satisfies Equation (7.6) and the asymptotic vari-
ance of LTS(τ ) is
F −1 (1/2+τ /2) 2
F −1 (1/2−τ /2)
w dF (w)
V (LT S(τ ), F ) = . (7.8)
[τ − 2F (1/2 + τ /2)f(F −1 (1/2 + τ /2))]2
−1
See Rousseeuw and Leroy (1987, p. 180, p. 191), and Tableman (1994a, p.
337).
b) The estimator β̂ LT A satisfies Equation (7.6) and the asymptotic vari-
ance of LTA(τ ) is
τ
V (LT A(τ ), F ) = . (7.9)
4[f(0) − f(F −1 (1/2 + τ /2))]2
See Tableman (1994b, p. 392) and Hössjer (1994).
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 221
τ σT2 F (−k, k)
V (LT S(τ ), F ) = (7.10)
[τ − 2kf(k)]2
and
τ
V (LT A(τ ), F ) = (7.11)
4[f(0) − f(k)]2
where
k = F −1 (0.5 + τ /2). (7.12)
2kφ(k)
σ 2[1 − ].
2Φ(k) − 1
while
τ 2πτ
V (LT A(τ ), Φ) = = (7.14)
4[φ(0) − φ(k)]2 4[1 − exp(−k 2 /2)]2
Thus for τ ≥ 1/2, LTS(τ ) has breakdown value of 1 − τ and Gaussian effi-
ciency
1
= τ − 2kφ(k). (7.15)
V (LT S(τ ), Φ)
The 50% breakdown estimator LTS(0.5) has a Gaussian efficiency of 7.1%.
If it is appropriate to reduce the amount of trimming, we can use the 25%
breakdown estimator LTS(0.75) which has a much higher Gaussian efficiency
of 27.6% as reported in Ruppert (1992, p. 255). Also see the column labeled
“Normal” in table 1 of Hössjer (1994).
Example 7.2: Double Exponential Errors. The double exponential
(Laplace) distribution is interesting since the L1 estimator corresponds to
maximum likelihood and so L1 beats OLS, reversing the comparison of the
normal case. For a double exponential DE(0, 1) random variable,
2 − (2 + 2k + k 2 ) exp(−k)
V (LT S(τ ), DE(0, 1)) =
[τ − k exp(−k)]2
while
τ 1
V (LT A(τ ), DE(0, 1)) = =
4[0.5 − 0.5 exp(−k)] 2 τ
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 223
where k = − log(1 − τ ). Note that LTA(0.5) and OLS have the same asymp-
totic efficiency at the double exponential distribution. Also see Tableman
(1994a,b).
Example 7.3: Cauchy Errors. Although the L1 estimator and the
trimmed estimators have finite variance when the errors are Cauchy, the
OLS estimator has infinite variance (because the Cauchy distribution has
infinite variance). If XT is a Cauchy C(0, 1) random variable symmetrically
truncated at −k and k, then
k − tan−1 (k)
VAR(XT ) = .
tan−1 (k)
Hence
2k − πτ
V (LT S(τ ), C(0, 1)) = 2k
π[τ − π(1+k 2) ]
2
and
τ
V (LT A(τ ), C(0, 1)) =
4[ π1 − 1
π(1+k2 )
]2
where k = tan(πτ /2). The LTA sampling variance converges to a finite value
as τ → 1 while that of LTS increases without bound. LTS(0.5) is slightly
more efficient than LTA(0.5), but LTA pulls ahead of LTS if the amount of
trimming is very small.
c
c
2 2
r(i) (β̂LT S ) ≤ r(i) (b)
i=1 i=1
where b is any p × 1 vector. Without loss of generality, assume that the cases
have been reordered so that the first c cases correspond to the cases with the
c smallest residuals. Let β̂ OLS (c) denote the OLS fit to these c cases. By the
definition of the OLS estimator,
c
c
ri2 (β̂OLS (c)) ≤ ri2 (b)
i=1 i=1
where b is any p × 1 vector. Hence β̂OLS (c) also minimizes the LTS criterion
and thus β̂ OLS (c) is an LTS estimator. The proofs of b) and c) are similar.
QED
Definition 7.7. In regression, an elemental set is a set of p cases.
One way to compute these estimators exactly is to generate all C(n, c)
subsets of size c, compute the classical estimator b on each subset, and find
the criterion Q(b). The robust estimator is equal to the bo that minimizes
the criterion. Since c ≈ n/2, this algorithm is impractical for all but the
smallest data sets. Since the L1 fit is an elemental fit, the LTA estimator can
be found by evaluating all C(n, p) elemental sets. See Hawkins and Olive
(1999b). Since any Chebyshev fit is also a Chebyshev fit to a set of p + 1
cases, the LQS estimator can be found by evaluating all C(n, p+1) cases. See
Stromberg (1993ab) and Appa and Land (1993). The LMS, LTA, and LTS
estimators can also be evaluated exactly using branch and bound algorithms
if the data set size is small enough. See Agulló (1997, 2001).
Typically HB algorithm estimators should not be used unless the criterion
complexity is O(n). The complexity of the estimator depends on how many
fits are computed and on the complexity of the criterion evaluation. For
example the LMS and LTA criteria have O(n) complexity while the depth
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 225
criterion complexity is O(np−1 log n). The LTA and depth estimators eval-
uates O(np ) elemental sets while LMS evaluates the O(np+1 ) subsets of size
p+1. The LQD criterion complexity is O(n2 ) and evaluates O(n2(p+1) ) subsets
of case distances.
Consider the algorithm that takes a subsample of nδ cases and then
computes the exact algorithm to this subsample. Then the complexities
of the LTA, LMS, depth and LQD algorithms are O(nδ(p+1) ), O(nδ(p+2) ),
O(nδ(2p−1) log nδ ) and O(nδ(2p+4) ), respectively. The convergence rates of the
estimators are nδ/3 for LMS and nδ/2 for the remaining √ three estimators (if
the LTA estimator does indeed have the conjectured n convergence rate).
These algorithms rapidly become impractical as n and p increase. For ex-
ample, if n = 100 and δ = 0.5, use p < 7, 6, 4, 2 for these LTA, LMS, depth,
and LQD algorithms respectively. If n = 10000, this LTA algorithm may not
be practical even for p = 3. These results suggest that the LTA and LMS
approximations will be more interesting than depth or LQD approximations
unless a computational breakthrough is made for the latter two estimators.
We simulated LTA and LTS for the location model using normal, dou-
ble exponential, and Cauchy error models. For the location model, these
estimators can be computed exactly: find the order statistics
of the data. For LTS compute the sample mean and for LTA compute the
sample median (or the low or high median) and evaluate the LTS and LTA
criteria of each of the n−c+1 “c-samples” Y(i) , . . . , Y(i+c−1) , for i = 1, . . . , n−
c + 1. The minimum across these samples then defines the LTA and LTS
estimates.
We computed the sample standard deviations of the resulting location es-
timate from 1000 runs of each sample size studied. The results are shown in
Table 7.1. For Gaussian errors, the observed standard deviations are smaller
than the asymptotic standard deviations but for the double exponential er-
rors, the sample size needs to be quite large before the observed standard
deviations agree with the asymptotic theory.
Table 7.2 presents the results of a small simulation study. We compared
ALTS(τ ) for τ = 0.5, 0.75, and 0.9 with RLTS(6) for 6 different error distribu-
tions – the normal(0,1), Laplace, uniform(−1, 1) and three 60% N(0,1) 40 %
contaminated normals. The three contamination scenarios were N(0,100) for
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 226
√
ter choice than the inconsistent estimators is to use the easily computed n
consistent HB CLTx estimators given in Theorem 8.7.) The concentration
algorithm used 300 starts for the location contamination distributions, and
50 starts for all others, preliminary experimentation having indicated that
this many starts were sufficient. Comparing the ‘conc’ mean squared errors
with the corresponding ‘elem’ confirms the recommendations in Hawkins and
Olive (2002) that far more than 3000 elemental starts are necessary to achieve
good results. The ‘elem’ runs also verify that second-stage refinement, as sup-
plied by the RLTS approach, is not sufficient to overcome the deficiencies in
the poor initial estimates provided by the raw elemental approach.
The RLTS estimator was, with one exception, either the best of the 4
estimators or barely distinguishable from the best. The single exception
was the concentration algorithm with the contaminated normal distribution
F (x) = 0.6Φ(x) + 0.4Φ(x − 5.5), where most of the time it covered all cases.
We already noted that location contamination with this mean and this choice
of k is about the worst possible for the RLTS estimator, so that this worst-
case performance is still about what is given by the more recent recommen-
dations for ALTx coverage – 75% or 90% is positive. This is reinforced by
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 228
150
50
OLS residuals
-50
-150
150
50
L1 residuals
-50
-150
50
0
ALMS residuals
-100
-200
150
50
ALTS residuals
-50
-150
0
-1000
MBA residuals
-3000
For a fixed xj consider the ordered distances D(1)(xj ), ..., D(n)(xj ). Next,
let β̂ j (α) denote the OLS fit to the min(p + 3 + αn/100
, n) cases with
the smallest distances where the approximate percentage of cases used is
α ∈ {1, 2.5, 5, 10, 20, 33, 50}. (Here x
is the greatest integer function so
7.7
= 7. The extra p + 3 cases are added so that OLS can be computed for
small n and α.) This yields seven OLS fits corresponding to the cases with
predictors closest to xj . A fixed number K of cases are selected at random
without replacement to use as the xj . Hence 7K OLS fits are generated. We
use K = 7 as the default. A robust criterion Q is used to evaluate the 7K
fits and the OLS fit to all of the data. Hence 7K + 1 OLS fits are generated
and the MBA estimator is the fit that minimizes the criterion. The median
squared residual, the LTA criterion, and the LATA criterion are good choices
for Q. Replacing the 7K + 1 OLS fits by L1 fits increases the resistance of
the MBA estimator.
Three ideas motivate this estimator. First, x-outliers, which are outliers
in the predictor space, tend to be much more destructive than Y -outliers
which are outliers in the response variable. Suppose that the proportion of
outliers is γ and that γ < 0.5. We would like the algorithm to have at least
one “center” xj that is not an outlier. The probability of drawing a center
that is not an outlier is approximately 1 − γ K > 0.99 for K ≥ 7 and this
result is free of p. Secondly, by using the different percentages of coverages,
for many data sets there will be a center and a coverage that contains no
outliers. √
Thirdly, the MBA estimator is a n consistent estimator. To see this,
assume that n is increasing to ∞. For each center xj,n there are 7 spheres
centered at xj,n . Let rj,h,n be the radius of the hth sphere with center xj,n .
Fix an extremely large N such that for n ≥ N these 7K regions in the
predictor space are fixed. Hence for n ≥ N the centers are xj,N and the
√ for j = 1, ..., K and h = 1, ..., 7. Since only a fixed number
radii are rj,h,N √
(7K + 1) of n consistent fits are computed, the final estimator is also a n
consistent estimator of β, regardless of how the final estimator is chosen (by
Pratt 1959).
Section 11.3 will compare the MBA estimator with other resistant es-
timators including the R/Splus estimator lmsreg and the trimmed views
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 231
7.7 Complements
The LTx and LATx estimators discussed in this chapter are not useful for
applications because they are impractical to compute; however, the criterion
are useful for making resistant or robust algorithm estimators. In particu-
lar the robust criterion are √
used in the MBA algorithm (see Problem 7.5)
and in the easily computed n consistent HB CLTx estimators described in
Theorem 8.7 and in Olive and Hawkins (2006).
Section 7.3 is based on Olive and Hawkins (1999) while Sections 7.2, 7.4,
7.5 and 7.6 follow Hawkins and Olive (1999b), Olive and Hawkins (2003) and
Olive (2005).
Several HB regression estimators are well known, and perhaps the first
proposed was the least median of squares (LMS) estimator. See Hampel
(1975, p. 380). For the location model, Yohai and Maronna (1976) and Butler
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 232
(1982) derived asymptotic theory for LTS. Rousseeuw (1984) generalized the
location LTS estimator to the LTS regression estimator and the minimum
covariance determinant estimator for multivariate location and dispersion
(see Chapter 10). Bassett (1991) suggested the LTA estimator for location
and Hössjer (1991) suggested the LTA regression estimator.
Two stage regression estimators compute a high breakdown regression
(or multivariate location and dispersion) estimator in the first stage. The
initial estimator is used to weight cases or as the initial estimator in a one
step Newton’s method procedure. The goal is for the two stage estimator
to inherit the outlier resistance properties of the initial estimator while hav-
ing high asymptotic efficiency when the errors follow a zero mean Gaussian
distribution. The theory for many of these estimators is often rigorous, but
the estimators are even less practical to compute than the initial estima-
tors. There are dozens of references including Jureckova and Portnoy (1987),
Simpson, Ruppert and Carroll (1992), Coakley and Hettmansperger (1993),
Chang, McKean, Naranjo and Sheather (1999), and He, Simpson and Wang
(2000).
The “cross checking estimator,” see He and Portnoy (1992, p. 2163) and
Davies (1993, p. 1981), computes a high breakdown estimator and OLS and
uses OLS if the two estimators are sufficiently close. The easily computed
HB estimators from Theorem 8.7 (and Olive and Hawkins 2006) make two
stage estimators such as the cross checking estimator practical for the first
time.
The theory of the RLTx estimator is very simple, but it can be used to
understand other results. For example, Theorem 7.3 will hold as long as
the initial estimator b used to compute Cn is consistent. In other words,
CLMS(0.5) (from Theorem 8.7) could be used as the√initial estimator for the
RLTS estimator. Suppose that the easily computed n consistent HB CLTS
estimator b (from Theorem 8.7) is used. If the CLTS(0.99) estimator does
indeed have high Gaussian efficiency, then the RLTS estimator that uses b as
the initial estimator will also have high Gaussian efficiency. Similar results
have appeared in the literature, but their proofs are very technical, often
requiring the theory of empirical processes.
The major drawback of high breakdown estimators that have nice the-
oretical results such as high efficiency is that they tend to be impractical
to compute. If an inconsistent zero breakdown initial estimator is used, as
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 233
in most of the literature and in the simulation study in Section 7.5, then
the final estimator (including even the simplest two stage estimators such
as the cross checking and RLTx estimators) also has zero breakdown and
√
is often inconsistent. Hence n consistent resistant estimators such as the
MBA estimator often have higher outlier resistance than zero breakdown
implementations of HB estimators such as ltsreg.
Another drawback of high breakdown estimators that have high efficiency
is that they tend to have considerably more bias than estimators such as
LTS(0.5) for many outlier configurations. For example the fifth row of Ta-
ble 7.2 shows that the RLTS estimator can perform much worse than the
ALTS(0.5) estimator if the outliers are within the k = 6 screen.
7.8 Problems
R/Splus Problems
Warning: Use the command source(“A:/rpack.txt”) to download
the programs. See Preface or Section 14.2. Typing the name of the
rpack function, eg mbamv, will display the code for the function. Use the
args command, eg args(mbamv), to display the needed arguments for the
function.
7.1. a) Download the R/Splus function nltv that computes the asymp-
totic variance of the LTS and LTA estimators if the errors are N(0,1).
b) Enter the commands nltv(0.5), nltv(0.75), nltv(0.9) and nltv(0.9999).
Write a table to compare the asymptotic variance of LTS and LTA at these
coverages. Does one estimator always have a smaller asymptotic variance?
7.2. a) Download the R/Splus function deltv that computes the asymp-
totic variance of the LTS and LTA estimators if the errors are double expo-
nential DE(0,1).
b) Enter the commands deltv(0.5), deltv(0.75), deltv(0.9) and deltv(0.9999).
Write a table to compare the asymptotic variance of LTS and LTA at these
coverages. Does one estimator always have a smaller asymptotic variance?
7.3. a) Download the R/Splus function cltv that computes the asymp-
totic variance of the LTS and LTA estimators if the errors are Cauchy C(0,1).
CHAPTER 7. ROBUST AND RESISTANT REGRESSION 234
humans but a few are based on apes. The MBA LATA estimator will often
give the cases corresponding to apes larger absolute residuals than the MBA
estimator based on MED(ri2 ).
e) Use the command mlrplot2(buxx,buxy) until the outliers are clustered
about the identity line in one of the two forward response plots. (This will
usually happen within 10 or fewer runs. Pressing the “up arrow” will bring
the previous command to the screen and save typing.) Then include the
resulting plot in Word. Which estimator went through the outliers and which
one gave zero weight to the outliers?
f) Use the command mlrplot2(hx,hy) several times. Usually both MBA es-
timators fail to find the outliers for this artificial Hawkins data set that is also
analyzed by Atkinson and Riani (2000, section 3.1). The lmsreg estimator
can be used to find the outliers. In Splus, use the command ffplot(hx,hy) and
in R use the commands library(lqs) and ffplot2(hx,hy). Include the resulting
plot in Word.
Chapter 8
Y = Xβ + e
denote the set of indices for the ith elemental set. Since there are n cases,
h1 , ..., hp are p distinct integers between 1 and n. For example, if n = 7 and
p = 3, the first elemental set may use cases J1 = {1, 7, 4}, and the second
elemental set may use cases J2 = {5, 3, 6}. The data for the ith elemental set
is (Y Jh , X Jh ) where Y Jh = (Yh1 , ..., Yhp)T is a p × 1 vector, and the p × p
236
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 237
matrix
xTh1 xh1,1 xh1,2 . . . xh1,p
xTh2 xh2,1 xh2,2 . . . xh2,p
X Jh = .. = .. .. .. .. .
. . . . .
xThp xhp,1 xhp,2 . . . xhp,p
bJh = X −1
Jh Y Jh
bA = argminh=1,...,Kn Q(bJh ).
j=1 i=1
From now on, unless otherwise stated, we will use the spectral norm as
the matrix norm and the Euclidean norm as the vector norm.
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 239
1 1
≤ ≤ X −1
J . (8.4)
p maxi,j |xhi,j | X J
for some real number M > 0 that does not depend on n. Then for any
elemental set X J ,
1
X −1
J ≥ . (8.5)
pM
QED
In proving consistency results, there is an infinite sequence of estimators
that depend on the sample size n. Hence the subscript n will be added to
the estimators. Refer to Remark 2.4 for the definition of convergence in
probability.
Definition 8.6. Lehmann (1999, p. 53-54): a) A sequence of random
variables Wn is tight or bounded in probability, written Wn = OP (1), if for
every
> 0 there exist positive constants D and N such that
P (|Wn | ≤ D ) ≥ 1 −
for all n ≥ N .
d) Similar notation is used for a k × r matrix A = [ai,j ] if each element
ai,j has the desired property. For example, A = OP (n−1/2 ) if each ai,j =
OP (n−1/2).
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 241
for some nondegenerate random variable X, then both Wn and β̂n have
convergence rate nδ .
If Wn has convergence rate nδ , then Wn has tightness rate nδ , and the term
“tightness” will often be omitted. Notice that if Wn P Xn , then Xn P Wn ,
Wn = OP (Xn ) and Xn = OP (Wn ). Notice that if Wn = OP (n−δ ), then nδ is a
lower bound on the rate of Wn . As an example, if LMS, OLS or L1 are used
for β̂, then Wn = OP (n−1/3 ), but Wn P n−1/3 for LMS while Wn P n−1/2
for OLS and L1 . Hence the rate for OLS and L1 is n1/2.
To examine the lack of consistency of the basic resampling algorithm
estimator bA,n meant to approximate the theoretical estimator β̂ Q,n , recall
that the key parameter of the basic resampling algorithm is the number of
elemental sets Kn ≡ K(n, p). Typically Kn is a fixed number, eg Kn ≡ K =
3000, that does not depend on n.
Example 8.2. This example illustrates the basic resampling algorithm
with Kn = 2. Let the data consist of the five (xi, yi ) pairs (0,1), (1,2), (2,3),
(3,4), and (1,11). Then p = 2 and n = 5. Suppose the criterion Q is the
median of the n squared residuals and that J1 = {1, 5}. Then observations
(0, 1) and (1, 11) were selected. Since bJ1 = (1, 10)T , the estimated line
is y = 1 + 10x, and the corresponding residuals are 0, −9, −18, −27, and
0. The criterion Q(bJ1 ) = 92 = 81 since the ordered squared residuals are
0, 0, 81, 182 , and 272 . If observations (0, 1) and (3, 4) are selected next, then
J2 = {1, 4}, bJ2 = (1, 1)T , and 4 of the residuals are zero. Thus Q(bJ2 ) = 0
and bA = bJ2 = (1, 1)T . Hence the algorithm produces the fit y = 1 + x.
Example 8.3. In the previous example the algorithm fit was reasonable,
but in general using a fixed Kn ≡ K in the algorithm produces inconsistent
estimators. To illustrate this claim, consider the location model Yi = β + ei
where the ei are iid and β is a scalar (since p = 1 in the location model). If β
was known, the natural criterion for an estimator bn of β would be Q(bn ) =
|bn − β|. For each sample size n, K elemental sets Jh,n = {hn }, h = 1, ..., K
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 242
of size p = 1 are drawn with replacement from the integers 1, ..., n. Denote
the resulting elemental fits by
bJh,n = Yhn
for h = 1, ..., K. Then the “best fit” Yo,n minimizes |Yhn − β|. If α > 0, then
provided that the errors have mass outside of [−α, α], and thus Yo,n is not
a consistent estimator. The inequality is needed since the Yhn may not be
distinct: the inequality could be replaced with equality if the Y1n , ..., YKn were
an iid sample of size K. Since α > 0 was arbitrary in the above example,
the inconsistency result holds unless the iid errors are degenerate at zero.
The basic idea is from sampling theory. A fixed finite sample can be used
to produce an estimator that contains useful information about a population
parameter, eg the population mean, but unless the sample size n increases to
∞, the confidence interval for the population parameter will have a length
bounded away from zero. In particular, if Y n (K) is a sequence of sample
means based on samples of size K = 100, then Y n (K) is not a consistent
estimator for the population mean.
The following notation is useful for the general regression setting and
will also be used for some algorithms that modify the basic resampling algo-
rithm. Let bsi,n be the ith elemental fit where i = 1, ..., Kn and let bA,n be the
algorithm estimator; that is, bA,n is equal to the bsi,n that minimized the cri-
terion Q. Let β̂ Q,n denote the estimator that the algorithm is approximating,
eg β̂ LT A,n . Let bos,n be the “best” of the K elemental fits in that
where the Euclidean norm is used. Since the algorithm estimator is an ele-
mental fit bsi,n ,
bA,n − β ≥ bos,n − β.
Thus an upper bound on the rate of bos,n is an upper bound on the rate of
bA,n .
Theorem 8.2. Let the number of randomly selected elemental sets
Kn → ∞ as n → ∞. Assume that the error distribution possesses a density
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 243
elemental sets generated. As an example where the elemental sets are not
chosen randomly, consider the L1 criterion. Since there is always an elemental
L1 fit, this fit has n1/2 convergence rate and is a consistent estimator of β.
Here we can take Kn ≡ 1, but the elemental set was not drawn randomly.
Using brain power to pick elemental sets is frequently a good idea.
1/p
It is also crucial to note that the Kn rate is only an upper bound on the
rate of the algorithm estimator bA,n . It is possible that the best elemental set
has a good convergence rate while the basic resampling algorithm estimator
is inconsistent.
Corollary 8.3. If the number Kn ≡ K of randomly selected elemental
sets is fixed and free of the sample size n, eg K = 3000, then the algorithm
estimator bA,n is an inconsistent estimator of β.
Conjecture 8.1. Suppose that the errors possess a density that is posi-
tive and continuous on the real line, that β̂Q,n − β = OP (n−1/2 ) and that
Kn ≤ C(n, p) randomly selected elemental sets are used in the algorithm.
−1/2p
Then the algorithm estimator satisfies bA,n − β = OP (Kn ).
Remark 8.2. This rate √ can be achieved if the algorithm minimizing Q
over all elemental subsets is n consistent (eg regression depth, see Bai and
He 1999). Randomly select g(n) cases and let Kn = C(g(n), p). Then apply
the all elemental subset algorithm to the g(n) cases. Notice that an upper
bound on the rate of bos,n is g(n) while
bA,n − β = OP ((g(n))−1/2).
Sometimes the notation bsi,n = b0i,n for the ith start and bai,n for the
ith attractor will be used. Using k = 10 concentration steps often works
well, and iterating until convergence is usually fast (in this case k = ki
depends on i). The “h–set” basic resampling algorithm uses starts that are
fits to randomly selected sets of h ≥ p cases, and is a special case of the
concentration algorithm with k = 0.
The notation CLTS, CLMS and CLTA will be used to denote concentra-
tion algorithms for LTA, LTS and LMS, respectively. Consider the LTS(cn )
criterion. Suppose the ordered squared residuals from the start b0k are ob-
tained. Then b1k is simply the OLS fit to the cases corresponding to the cn
smallest squared residuals. Denote these cases by i1 , ..., icn . Then
cn
cn
cn
cn
2
r(i) (b1k ) ≤ ri2j (b1k ) ≤ ri2j (b0k ) = 2
r(i) (b0k )
i=1 j=1 j=1 j=1
where the second inequality follows from the definition of the OLS estimator.
Convergence to the attractor tends to occur in a few steps.
A simplified version of the CLTS(c) algorithms of Ruppert (1992), Vı́šek
(1996), Hawkins and Olive (1999a) and Rousseeuw and Van Driessen (2000,
2002) uses Kn elemental starts. The LTS(c) criterion is
c
2
QLT S (b) = r(i) (b) (8.8)
i=1
2
where r(i) (b) is the ith smallest squared residual. For each elemental start
find the exact-fit bsj to the p cases in the elemental start and then get the
c smallest squared residuals. Find the OLS fit to these c cases and find
the resulting c smallest squared residuals, and iterate for k steps (or until
convergence). Doing this for Kn elemental starts leads to Kn (not necessarily
distinct) attractors baj . The algorithm estimator β̂ALT S is the attractor that
minimizes Q. Substituting the L1 or Chebyshev fits and LTA or LMS criteria
for OLS in the concentration step leads to the CLTA or CLMS algorithm.
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 246
a) A Start for the Animal Data b) The Attractor for the Start
8
8
6
6
4
4
Y
Y
2
2
0
0 5 10 0 5 10
X X
Figure 8.1: The Highlighted Points are More Concentrated about the At-
tractor
Figure 8.1a shows the scatterplot of x and y. The start is also shown and
the 14 cases corresponding to the smallest absolute residuals are highlighted.
The L1 fit to these c highlighted cases is b2,1 = (2.076, 0.979)T and
14
|r|(i)(b2,1) = 6.990.
i=1
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 247
8
6
6
4
4
Y
Y
2
2
0
0 5 10 0 5 10
X X
Figure 8.1b shows the attractor and that the c highlighted cases correspond-
ing to the smallest absolute residuals are much more concentrated than those
in Figure 8.1a. Figure 8.2a shows 5 randomly selected starts while Figure
8.2b shows the corresponding attractors. Notice that the elemental starts
have more variablity than the attractors, but if the start passes through an
outlier, so does the attractor.
Notation for the attractor needs to be added to the notation used for the
basic resampling algorithm. Let bsi,n be the ith start, and let bai,n be the
ith attractor. Let bA,n be the algorithm estimator, that is, the attractor that
minimized the criterion Q. Let β̂Q,n denote the estimator that the algorithm
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 248
Since β̂ and Y have the same characteristic functions, they have the same
distribution. Thus β̂2 has the same distribution as
QED
This result shows the inadequacy of elemental sets in high dimensions.
For a trial fit to provide a useful preliminary classification of cases into inliers
and outliers requires that it give a reasonably precise slope. However if p is
large, this is most unlikely; the density of (b−β)T (b−β) varies near zero like
p
[(b − β)T (b − β)]( 2 −1) . For moderate to large p, this implies that good trial
slopes will be extremely uncommon and so enormous numbers of random
elemental sets will have to be generated to have some chance of finding one
that gives a usefully precise slope estimate. The only way to mitigate this
effect of basic resampling is to use larger values of h, but this negates the
main virtue elemental sets have, which is that when outliers are present, the
smaller the h the greater the chance that the random subset will be clean.
The following two propositions examine increasing the start size. The
first result (compare Remark 8.3) proves that increasing the start size from
elemental to h ≥ p results in a zero breakdown inconsistent estimator. Let
the k–step CLTS estimator be the concentration algorithm estimator for LTS
that uses k concentration steps. Assume that the number of concentration
steps k and the number of starts Kn ≡ K do not depend on n (eg k = 10
and K = 3000, breakdown is defined in Section 9.4).
Proposition 8.4. Suppose that each start uses h randomly selected cases
and that Kn ≡ K starts are used. Then
i) the (“h-set”) basic resampling estimator is inconsistent.
ii) The k–step CLTS estimator is inconsistent.
iii) The breakdown value is bounded above by K/n.
Proof. To prove i) and ii), notice that each start is inconsistent. Hence
each attractor is inconsistent by He and Portnoy (1992). Choosing from K
inconsistent estimators still results in an inconsistent estimator. To prove iii)
replace one observation in each start by a high leverage case (with y tending
to ∞). QED
The next result shows that the situation changes dramatically if K is
fixed but the start size h = hn = g(n) where g(n) → ∞. In particular,
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 251
if several starts with rate n1/2 are used, the final estimator also has rate
n1/2. The drawback to these algorithms is that they may not have enough
outlier resistance. Notice that the basic resampling result below is free of the
criterion. Suppose that β̂ 1, ..., β̂K are consistent estimators of β each with
the same rate g(n). Pratt (1959) shows that if β̂ A is an estimator obtained
by choosing one of the K estimators, then β̂ A is a consistent estimator of β
with rate g(n).
Proposition 8.5. Suppose Kn ≡ K starts are used and that all starts
have subset size hn = g(n) ↑ ∞ as n → ∞. Assume that the estimator
applied to the subset has rate nδ .
i) For the hn -set basic resampling algorithm, the algorithm estimator has
rate [g(n)]δ .
ii) Under mild regularity conditions (eg given by He and Portnoy 1992), the
k–step CLTS estimator has rate [g(n)]δ .
Proof. i) The hn = g(n) cases are randomly sampled without replace-
ment. Hence the classical estimator applied to these g(n) cases has rate
[g(n)]δ . Thus all K starts have rate [g(n)]δ , and the result follows by Pratt
(1959). ii) By He and Portnoy (1992), all K attractors have [g(n)]δ rate, and
the result follows by Pratt (1959). QED
These results show that fixed Kn ≡ K elemental methods are inconsis-
tent. Several simulation studies have shown that the versions of the resam-
pling algorithm that use a fixed number of elemental starts provide fits with
√
behavior that conforms with the asymptotic behavior of the n consistent
target estimator. These paradoxical studies can be explained by the following
proposition (a recasting of a coupon collection problem).
Proposition 8.6. Suppose that Kn ≡ K random starts of size h are
selected and let Q(1) ≤ Q(2) ≤ · · · ≤ Q(B) correspond to the order statistics
of the criterion values of the B = C(n, h) possible starts of size h. Let R be
the rank of the smallest criterion value from the K starts. If P (R ≤ Rα ) = α,
then
Rα ≈ B[1 − (1 − α)1/K ].
Proof. If Wi is the rank of the ith start, then W1 , ..., WK are iid discrete
uniform on {1, ..., B} and R = min(W1 , ..., WK ). If r is an integer in [1, B],
then
B−r K
P (R ≤ r) = 1 − ( ) .
B
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 252
B KB2 B2
E(R) ≈ 1 + , and VAR(R) ≈ ≈ .
K +1 (K + 1)2 (K + 2) K2
Notice that bo,n is not an estimator since β is unknown, but since the algo-
rithm estimator is an elemental fit, bA,n − β ≥ bo,n − β, and an upper
bound on the rate of bo,n is an upper bound on the rate of bA,n . Theorem 8.2
1/p
showed that the rate of the bo,n ≤ Kn , regardless of the criterion Q. This
result is one of the most powerful tools for examining the behavior of robust
estimators actually used in practice. For example, many basic resampling
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 254
algorithms use Kn = O(n) elemental sets drawn with replacement from all
C(n, p) elemental sets. Hence the algorithm estimator bA,n has a rate ≤ n1/p.
1/p
This section will show that the rate of bo,n is Kn and suggests that the
number of elemental sets bi,n that satisfy bi,n − β ≤ Mnδ (where M > 0
is some constant and 0 < δ ≤ 1) is proportional to np(1−δ) .
Two assumptions are used.
(A1) The errors are iid, independent of the predictors, and have a density f
that is positive and continuous in a neighborhood of zero.
(A2) Let τ be proportion of elemental sets J that satisfy X −1 J ≤ B for
some constant B > 0. Assume τ > 0.
These assumptions are reasonable, but results that do not use (A2) are
given later. If the errors can be arbitrarily placed, then they could cause the
estimator to oscillate about β. Hence no estimator would be consistent for
β. Note that if
> 0 is small enough, then P (|ei | ≤
) ≈ 2
f(0). Equations
(8.2) and (8.3) suggest that (A2) will hold unless the data is such that nearly
all of the elemental trial designs X J have badly behaved singular values.
Theorem 8.8. Assume that all C(n, p) elemental subsets are searched
and that (A1) and (A2) hold. Then bo,n − β = OP (n−1 ).
Proof. Let the random variable Wn, count the number of errors ei that
satisfy |ei | ≤ M /n for i = 1, ..., n. For fixed n, Wn, is a binomial random
variable with parameters n and Pn where nPn → 2f(0)M as n → ∞. Hence
Wn, converges in distribution to a Poisson(2f(0)M ) random variable, and
for any fixed integer k > p, P (Wn, > k) → 1 as M → ∞ and n → ∞. Hence
if n is large enough, then with arbitrarily high probability there exists an M
such that at least C(k, p) elemental sets Jhn have all |ehn i | ≤ M /n where
the subscript hn indicates that the sets depend on n. By condition (A2),
√
the proportion of these C(k, p) fits that satisfy bJhn − β ≤ B pM /n is
greater than τ. If k is chosen sufficiently large, and if n is sufficiently large,
√
then with arbitrarily high probability, bo,n − β ≤ B pM /n and the result
follows. QED
Corollary 8.9. Assume that Hn ≤ n but Hn ↑ ∞ as n → ∞. If (A1)
and (A2) hold, and if Kn = Hnp randomly chosen elemental sets are used,
−1/p
then bo,n − β = OP (Hn−1 ) = OP (Kn ).
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 255
Proof. Suppose Hn cases are drawn without replacement and all C(Hn , p)
∝ Hnp elemental sets are examined. Then by Theorem 8.8, the best elemental
set selected by this procedure has rate Hn . Hence if Kn = Hnp randomly cho-
sen elemental sets are used and if n is sufficiently large, then the probability
of drawing an elemental set Jhn such that bJhn − β ≤ M Hn−1 goes to one
as M → ∞ and the result follows. QED
Suppose that an elemental set J is “good” if bJ − β ≤ M Hn−1 for some
constant M > 0. The preceding proof used the fact that with high probabil-
ity, good elemental sets can be found by a specific algorithm that searches
Kn ∝ Hnp distinct elemental sets. Since the total number of elemental sets
is proportional to np , an algorithm that randomly chooses Hnp elemental sets
will find good elemental sets with arbitrarily high probability. For example,
the elemental sets could be drawn with or without replacement from all of
the elemental sets. As another example, draw a random permutation of the
n cases. Let the first p cases be the 1st elemental set, the next p cases the
2nd elemental set, etc. Then about n/p elemental sets are generated, and
the rate of the best elemental set is n1/p.
Also note that the number of good sets is proportional to np Hn−p . In
particular, if Hn = nδ where 0 < δ ≤ 1, then the number of “good” sets
is proportional to np(1−δ). If the number of randomly drawn elemental sets
Kn = o((Hn )p ), then bA,n −β = OP (Hn−1 ) since P (bo,n −β ≤ Hn−1 M ) →
0 for any M > 0.
A key assumption to Corollary 8.9 is that the elemental sets are randomly
drawn. If this assumption is violated, then the rate of the best elemental set
could be much higher. For example, the single elemental fit corresponding
to the L1 estimator could be used, and this fit has a n1/2 rate.
The following argument shows that similar results hold if the predictors
are iid with a multivariate density that is everywhere positive. For now,
assume that the regression model contains a constant: x = (1, x2 , ..., xp)T .
Construct a (hyper) pyramid and place the “corners” of the pyramid into
a p × p matrix W . The pyramid defines p “corner regions” R1 , ..., Rp. The
p points that form W are not actual observations, but the fit bJ can be
evaluated on W . Define the p × 1 vector z = W β. Then β = W −1 z, and
ẑ = W bJ is the fitted hyperplane evaluated at the corners of the pyramid.
If an elemental set has one observation in each corner region and if all p
absolute errors are less than
, then the absolute deviation |δi| = |zi − ẑi| <
,
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 256
i = 1, ..., p.
To fix ideas and notation, we will present three examples. The first two
examples consider the simple linear regression model with one predictor and
an intercept while the third example considers the multiple regression model
with two predictors and an intercept.
Example 8.5. Suppose the design has exactly two distinct predictor
values, (1, x1,2) and (1, x2,2), where x1,2 < x2,2 and
Notice that
β = X −1 z
where
z = (z1, z2 )T = (β1 + β2x1,2, β1 + β2 x2,2)T
and
1 x1,2
X= .
1 x2,2
If we assume that the errors are iid N(0, 1), then P (Yi = zj ) = 0 for j = 1, 2
and n ≥ 1. However,
Suppose that the elemental set J = {i1, i2} is such that xij = xj and |yij −
zj | <
for j = 1, 2. Then bJ = X −1 Y J and
√
bJ − β ≤ X −1 Y J − z ≤ X −1 2
.
ẑ = W bJ
is the fitted line evaluated at w1 and w2. Let the deviation vector
δ J = (δJ,1 , δJ,2)T
where
δ J,i = zi − ẑi .
Hence
bJ = W −1 (z − δ J )
and
|δJ,i | ≤
by construction. Thus
bJ − β = W −1 z − W −1 δ J − W −1 z
√
≤ W −1 δJ ≤ W −1 2
.
The basic idea is that if a fit is determined by one point from each region
and if the fit is good, then the fit has small deviation at points w1 and w2
because lines can’t bend. See Figure 8.3. Note that the bound is true for
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 258
•
••
••
•••
•
••
••
• ••
••
••
•••
•
••
••
• ••
••
••
•••
••
••
•••
••
••
••
••
• ••
• ••
••
••
•••
•
••
••
• ••
••
••
•••
•
••
••
••
••
every fit such that one point is in each region and both absolute errors are
less than
. The number of such fits can be enormous. For example, if
is a
constant, then the number of observations in region Ri with errors less than
is proportional to n for i = 1, 2. Hence the number of “good” fits from the
two regions is proportional to n2 .
Example 8.7. Now assume that p = 3 and Yi = β1 + β2xi,2 + β3xi,3 + ei
where the predictors (xi,2, xi,3 ) are scattered about the origin, eg iid N(0, I 2).
Now we need a matrix W and three regions with many observations that
have small errors. Let
1 a −a/2
W = 1 −a −a/2
1 0 a/2
for some a > 0 (eg a = 1). Note that the three points (a, −a/2)T , (−a, −a/2)T ,
and (0, a/2)T determine a triangle. Use this triangle as the pyramid. Then
the corner regions are formed by extending the three lines that form the
triangle and using points that fall opposite of a corner of the triangle. Hence
R1 = {(x2, x3)T : x3 < −a/2 and x2 > a/2 − x3},
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 259
bJ − β = W −1 z − W −1 δ J − W −1 z
√
≤ W −1 δJ ≤ W −1 3
.
For Example 8.7, there is a prism shaped region centered at the triangle
determined by W . Any elemental subset J with one point in each corner
region and with each absolute error less than
produces a plane that cuts
the prism. Hence each absolute deviation at the corners of the triangle is less
than
.
The geometry in higher dimensions uses hyperpyramids and hyperprisms.
When p = 4, the p = 4 rows that form W determine a 3–dimensional
pyramid. Again we have 4 corner regions and only consider elemental subsets
consisting of one point from each region with absolute errors less than
.
The resulting hyperplane will cut the hyperprism formed by extending the
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 260
Corner Region R3
3
2
1
x3
−3 −2 −1 0 1 2 3
x2
Figure 8.4: The Corner Regions for Two Predictors and a Constant.
Hence the probability that a randomly selected elemental set bJ that satisfies
√
bJ − β ≤ W −1 p M /Hn is bounded below by a probability that is
proportional to (M /Hn )p . If the number of randomly selected elemental sets
Kn = Hnp , then
√ M
P (bo,n − β ≤ W −1 p )→1
Hn
as M → ∞. Notice that one way to choose Kn is to draw Hn ≤ n cases
without replacement and then examine all Kn = C(Hn , p) elemental sets.
These remarks prove the following corollary.
Corollary 8.11. Assume that (A1) and (A3) hold. Let Hn ≤ n and
assume that Hn ↑ ∞ as n → ∞. If Kn = Hnp elemental sets are randomly
chosen then
bo,n − β = OP (Hn−1 ) = OP (Kn−1/p ).
In particular, if all C(n, p) elemental sets are examined, then bo,n −β =
OP (n−1 ). Note that Corollary 8.11 holds as long as the bulk of the data
satisfies (A1) and (A3). Hence if a fixed percentage of outliers are added to
clean cases, rather than replacing clean cases, then Corollary 8.11 still holds.
The following result shows that elemental fits can be used to approximate
any p × 1 vector c. Of course this result is asymptotic, and some vectors will
not be well approximated for reasonable sample sizes.
Theorem 8.12. Assume that (A1) and (A3) hold and that the error
density f is positive and continuous everywhere. Then the closest elemental
fit bc,n to any p × 1 vector c satisfies bc,n − c = OP (n−1 ).
Proof sketch. The proof is essentially the same. Sandwich the plane
determined by c by only considering points such that |gi | ≡ |yi − xTi c| < α.
Since the ei ’s have positive density, P (|gi | < α) ∝ 1/α) (at least for xi in
some ball of possibly huge radius R about the origin). Also the pyramid needs
to lie on the c-plane and the corner regions will have smaller probabilities.
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 263
8.4 Complements
The widely used basic resampling and concentration algorithms that use
a fixed number K of randomly drawn elemental sets are inconsistent, but
Theorem 8.7 shows that it is easy to modify some
√ of these algorithms so that
the easily computed modified estimator is a n consistent high breakdown
(HB) estimator. The√basic idea is to evaluate the criterion on K elemental
sets as well as on a n consistent estimator such as OLS and on an easily
computed HB but biased √estimator such as β̂ k,B . Similar ideas will be used to
create easily computed n consistent HB estimators of multivariate location
and dispersion. See Section 10.7.
The first two sections of this chapter followed Hawkins and Olive (2002)
and Olive and Hawkins (2006) closely. The “basic resampling”, or “ele-
mental set” method was used for finding outliers in the regression setting by
Siegel (1982), Rousseeuw (1984), and Hawkins, Bradu and Kass (1984). Fare-
brother (1997) sketches the history of elemental set methods. Also see Mayo
and Gray (1997). Hinich and Talwar (1975) used nonoverlapping elemental
sets as an alternative to least squares. Rubin (1980) used elemental sets for
diagnostic purposes. The “concentration” technique may have been intro-
duced by Devlin, Gnanadesikan and Kettenring (1975) although a similar
idea appears Gnanadesikan and Kettenring (1972, p. 94). The concentration
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 264
technique for regression was used by Ruppert (1992) and Hawkins and Olive
(1999a).
A different generalization of the elemental set method uses for its starts
subsets of size greater than p (Atkinson and Weisberg 1991). Another possi-
ble refinement is a preliminary partitioning of the cases (Woodruff and Rocke,
1994, Rocke, 1998, Rousseeuw and Van Driessen, 1999, 2002).
If an exact algorithm exists but an approximate algorithm is also used,
the two estimators should be distinguished in some manner. For example
β̂LM S could denote the estimator from the exact algorithm while β̂ALM S could
denote the estimator from the approximate algorithm. In the literature this
distinction is too seldomly made, but there are a few outliers. Portnoy (1987)
makes a distinction between LMS and PROGRESS LMS while Cook and
Hawkins (1990, p. 640) point out that the AMVE is not the minimum
volume ellipsoid (MVE) estimator (which is a high breakdown estimator of
multivariate location and dispersion that is sometimes used to define weights
in regression algorithms). Rousseeuw and Bassett (1991) find the breakdown
point and equivariance properties of the LMS algorithm that searches all
C(n, p) elemental sets. Woodruff and Rocke (1994, p. 889) point out that in
practice the algorithm is the estimator. Hawkins (1993a) has some results
when the fits are computed from disjoint elemental sets, and Rousseeuw
(1993, p. 126) states that the all subsets version of PROGRESS is a high
breakdown algorithm, but the random sampling versions of PROGRESS are
not high breakdown algorithms.
Algorithms which use one interchange on elemental sets may be compet-
itive. Heuristically, only p − 1 of the observations in the elemental set need
small absolute errors since the best interchange would be with the observa-
tion in the set with a large error and an observation outside of the set with a
very small absolute error. Hence K ∝ nδ(p−1) starts are needed. Since finding
the best interchange requires p(n − p) comparisons, the run time should be
competitive with the concentration algorithm. Another idea is to repeat the
interchange step until convergence. We do not know how many starts are
needed for this algorithm to produce good results.
Theorems 8.2 and 8.8 are an extension of Hawkins (1993a, p. 582) which
states that if the algorithm uses O(n) elemental sets, then at least one ele-
mental set b is likely to have its jth component bj close to the jth component
βj of β.
Note that one-step estimators can improve the rate of the initial estima-
tor. See for example Chang, McKean, Naranjo, and Sheather (1999) and
CHAPTER 8. ROBUST REGRESSION ALGORITHMS 265
Simpson, Ruppert, and Carroll (1992). Although the theory for the estima-
tors in these two papers requires an initial high breakdown estimator with
at least an n1/4 rate of convergence, implementations often use an initial
inconsistent, low breakdown algorithm estimator. Instead of using lmsreg
or ltsreg as the initial estimator, use the CLTS estimator of Theorem 8.7
(or the MBA or trimmed views estimators of Sections 7.6 and 10.5). The
CLTS estimator can also be used to create an asymptotically efficient high
breakdown cross checking estimator.
The Rousseeuw and Leroy (1987) data sets are available from the follow-
ing website
(http://www.uni-koeln.de/themen/Statistik/data/rousseeuw/).
(http://www.agoras.ua.ac.be/) and
(http://www.stat.umn.edu/ARCHIVES/archives.html).
8.5 Problems
8.1. Since an elemental fit b passes through the p cases, a necessary condition
for b to approximate β well is that all p errors be small. Hence no “good”
approximations will be lost when we consider only the cases with |ei | <
. If
the errors are iid, then for small
> 0, case i has
Hence if
= 1/n(1−δ) , where 0 ≤ δ < 1, find how many cases have small
errors.
8.2 Suppose that e1, ..., e100 are iid and that α > 0. Show that
Splus Problems
For problems 8.3 and 8.4, if the animal or Belgian telephone data sets
(Rousseeuw and Leroy 1987) are not available, use the following commands.
8.3. a) Download the Splus function conc2. This function does not work
in R.
b) Include the output from the following command in Word.
conc2(zx,zy)
8.4. a) Download the Splus function attract that was used to produce
Figure 8.2. This function does not work in R.
b) Repeat the following command five times.
> attract(zx,zy)
267
CHAPTER 9. RESISTANCE AND EQUIVARIANCE 268
the paradigm gives a useful initial model for the data. The assumption is
very widely used in the literature for diagnostics and robust statistics.
Remark 9.1. Suppose that the data set contains n cases with d outliers
and n − d clean cases. Suppose that h ≥ p cases are selected at random
without replacement. Let W count the number of the h cases that were
outliers. Then W is a hypergeometric(d, n − d, h) random variable and
&d'&n−d'
h j
P (W = j) = &n' ≈
j h−j
γ (1 − γ)h−j
h
j
where the contamination proportion γ = d/n and the binomial(h, ρ ≡ γ =
d/n) approximation to the hypergeometric(d, n − d, h) distribution is used.
In particular, the probability that the subset of h cases is clean = P (W =
0) ≈ (1 − γ)h which is maximized by h = p. Hence using elemental sets
maximizes the probability of getting a clean subset. Moreover, computing
the elemental fit is faster than computing the fit from h > p cases.
Remark 9.2. Now suppose that K elemental sets are chosen with re-
placement. If Wi is the number of outliers in the ith elemental set, then
the Wi are iid hypergeometric(d, n − d, p) random variables. Suppose that
it is desired to find K such that the probability P(that at least one of the
elemental sets is clean) ≡ P1 ≈ 1 − α where α = 0.05 is a common choice.
Then P1 = 1− P(none of the K elemental sets is clean)
≈ 1 − [1 − (1 − γ)p]K
by independence. Hence
α ≈ [1 − (1 − γ)p ]K
or
log(α) log(α)
K≈ ≈ (9.1)
log([1 − (1 − γ) ])
p −(1 − γ)p
using the approximation log(1 − x) ≈ −x for small x. Since log(.05) ≈ −3,
if α = 0.05, then
3
K≈ .
(1 − γ)p
Frequently a clean subset is wanted even if the contamination proportion
γ ≈ 0.5. Then for a 95% chance of obtaining at least one clean elemental set,
K ≈ 3 (2p ) elemental sets need to be drawn.
CHAPTER 9. RESISTANCE AND EQUIVARIANCE 269
K
γ 500 3000 10000 105 106 107 108 109
0.01 509 687 807 1036 1265 1494 1723 1952
0.05 99 134 158 203 247 292 337 382
0.10 48 65 76 98 120 142 164 186
0.15 31 42 49 64 78 92 106 120
0.20 22 30 36 46 56 67 77 87
0.25 17 24 28 36 44 52 60 68
0.30 14 19 22 29 35 42 48 55
0.35 11 16 18 24 29 34 40 45
0.40 10 13 15 20 24 29 33 38
0.45 8 11 13 17 21 25 28 32
0.50 7 9 11 15 18 21 24 28
Table 9.1 shows the largest value of p such that there is a 95% chance
that at least one of K subsamples is clean using the approximation given
by Equation (9.2). Hence if p = 28, even with one billion subsamples, there
is a 5% chance that none of the subsamples will be clean if the contamina-
tion proportion γ = 0.5. Since clean elemental fits have great variability, an
algorithm needs to produce many clean fits in order for the best fit to be
good. When contamination is present, all K elemental sets could contain
outliers. Hence basic resampling and concentration algorithms that only use
K elemental starts are doomed to fail if γ and p are large.
Remark 9.3: Breakdown. The breakdown value of concentration al-
gorithms that use K elemental starts is bounded above by K/n. (See Section
9.4 for more information about breakdown.) For example if 500 starts are
used and n = 50000, then the breakdown value is at most 1%. To cause a
regression algorithm to break down, simply contaminate one observation in
each starting elemental set so as to displace the fitted coefficient vector by a
large amount. Since K elemental starts are used, at most K points need to
be contaminated.
This is a worst-case model, but sobering results on the outlier resistance of
such algorithms for a fixed data set with d gross outliers can also be derived.
Assume that the LTA(c), LTS(c), or LMS(c) algorithm is applied to a fixed
data set of size n where n − d of the cases follow a well behaved model and
d < n/2 of the cases are gross outliers. If d > n − c, then every criterion
evaluation will use outliers, and every attractor will produce a bad fit even
if some of the starts are good. If d < n − c and if the outliers are far enough
from the remaining cases, then clean starts of size h ≥ p may result in clean
attractors that could detect certain types of outliers (that may need to be
hugely discrepant on the response scale).
Proposition 9.1. Let γo be the highest percentage of massive outliers
that a resampling algorithm can detect reliably. Then
n−c
γo ≈ min( , 1 − [1 − (0.2)1/K ]1/h)100% (9.3)
n
if n is large.
Proof. In Remark 9.2, change p to h to show that if the contamination
proportion γ is fixed, then the probability of obtaining at least one clean
subset of size h with high probability (say 1 − α = 0.8) is given by 0.8 =
CHAPTER 9. RESISTANCE AND EQUIVARIANCE 271
1 − [1 − (1 − γ)h ]K . Fix the number of starts K and solve this equation for
γ. QED
The value of γo depends on c > n/2 and h. To maximize γo , take c ≈ n/2
and h = p. For example, with K = 500 starts, n > 100, and h = p ≤ 20 the
resampling algorithm should be able to detect up to 24% outliers provided
every clean start is able to at least partially separate inliers from outliers.
However if h = p = 50, this proportion drops to 11%.
Remark 9.4: Hybrid Algorithms. More sophisticated algorithms use
both concentration and partitioning. Partitioning evaluates the start on a
subset of the data, and poor starts are discarded. This technique speeds
up the algorithm, but the consistency and outlier resistance still depends on
the number of starts. For example, Equation (9.3) agrees very well with the
Rousseeuw and Van Driessen (1999) simulation performed on a hybrid MCD
algorithm. (See Section 10.6.)
λbb + (1 − λ)b
where b is the fit from the current subsample and λ is between 0 and 1. Using
λ ≈ 0.9 may make sense. If the algorithm produces a good fit at some stage,
then many good fits will be examined with this technique.
Y = Y (X, Y ) = X β(X,
Y ), (9.5)
r = r(X, Y ) = Y − Y . (9.6)
is
Regression Equivariance: Let u be any p × 1 vector. Then β
regression equivariant if
β(X,
Y + Xu) = T (X, Y + Xu) = T (X, Y ) + u = β(X, Y ) + u. (9.7)
and
= r(X, Y ).
r(W , Z) = Z − Z
Note that the residuals are invariant under this type of transformation, and
note that if
u = −β,
then regression equivariance implies that we should not find any linear struc-
ture if we regress the residuals on X.
is scale equivariant if
Scale Equivariance: Let c be any scalar. Then β
β(X,
cY ) = T (X, cY ) = cT (X, Y ) = cβ(X, Y ). (9.8)
and
r(X, cY ) = c r(X, Y ).
Scale equivariance implies that if the Yi ’s are stretched, then the fits and the
residuals should be stretched by the same factor.
Affine Equivariance: Let A be any p × p nonsingular matrix. Then β
is affine equivariant if
β(XA,
Y ) = T (XA, Y ) = A−1 T (X, Y ) = A−1 β(X, Y ). (9.9)
Z
= W β(XA,
Y ) = XAA−1 β(X, Y ) = Y ,
and
= Y − Y = r(X, Y ).
r(XA, Y ) = Z − Z
CHAPTER 9. RESISTANCE AND EQUIVARIANCE 275
Note that both the predicted values and the residuals are invariant under an
affine transformation of the independent variables.
Permutation Invariance: Let P be an n×n permutation matrix. Then
P P = P P T = I n where I n is an n × n identity matrix and the superscript
T
bJ ≤ X −1
J Y J ,
and since x-outliers make X J large, x-outliers tend to drive bJ towards
zero not towards ∞. The x-outliers may make bJ large if they can make
the trial design X J nearly singular. Notice that Euclidean norm bJ can
easily be made large if one or more of the elemental response variables is
driven far away from zero.
CHAPTER 9. RESISTANCE AND EQUIVARIANCE 277
Example 9.2. Without loss of generality, assume that the good y’s are
contained in an interval [a, f] for some a and f. Assume that the regression
model contains an intercept β1. Then there exists an estimator bo of β such
that bo ≤ max(|a|, |f|) if d < n/2.
Proof. Let med(n) = med(y1 , ..., yn) and mad(n) = mad(y1 , ..., yn). Take
bo = (med(n), 0, ..., 0)T . Then bo = |med(n)| ≤ max(|a|, |f|). Note that
the median absolute residual for the fit bo is equal to the median absolute
deviation mad(n) = med(|yi −med(n)|, i = 1, ..., n) ≤ f −a if d < (n+1)/2
.
QED
A high breakdown regression estimator is an estimator which has a bound-
ed median absolute residual even when close to half of the observations are
arbitrary. Rousseeuw and Leroy (1987, p. 29, 206) conjecture that high
breakdown regression estimators can not be computed cheaply, and they
conjecture that if the algorithm is also affine equivariant, then the complexity
of the algorithm must be at least O(np ). The following counterexample shows
that these two conjectures are false.
Example 9.3. If the model has an intercept, then an affine equivariant
high breakdown estimator β̂W LS (k) can be found by computing OLS to the
set of cases that have yi ∈ [med(y1, ..., yn) ± k mad(y1, ..., yn)] where k ≥ 1
(so at least half of the cases are used). When k = 1, this estimator uses the
“half set” of cases closest to med(y1, ..., yn). Hence 0.99β̂ W LS (1) could be
used as the biased HB estimator needed in Theorem 8.7.
which yields the predicted values Ŷi ≡ MED(n). The squared residual
Thus
√
MED(|r1(β̂ W LS )|, ..., |rn(β̂W LS )|) ≤ nj k MAD(n) < ∞.
Thus
Q(|cr1,1|, . . . , |crn,1 |) = |c|d Q(|r1,1|, . . . .|rn,1|) <
|c|d Q(|r1,2|, . . . , |rn,2|) = Q(|cr1,2|, . . . , |crn,2|),
and TQ is scale equivariant. QED
Since least squares is regression, scale, and affine equivariant, the fit from
an elemental or subset refinement algorithm that uses OLS also has these
properties provided that the criterion Q satisfies the condition in Lemma
9.2. If
Q = med(ri2 ),
then d = 2. If
h
Q= (|r|(i) )τ
i=1
or
n
Q= wi |ri |τ
i=1
then d = τ .
Corollary 9.3. Any low breakdown affine equivariant estimator can be
approximated by a high breakdown affine equivariant estimator.
Proof. Let β̂W LS (k2 ) be the estimator in Example 9.3 with k = k2 ≥ 1.
Let β̂ be the low breakdown estimator, and let
β̂approx = β̂ W LS (k2 ),
otherwise. If k1 > 1 is large, the approximation will be good. QED
Remark 9.5. Similar breakdown results hold for multivariate location
and dispersion estimators. See Section 10.5.
Remark 9.6. There are several important points about breakdown that
do not seem to be well known. First, a breakdown result is weaker than even
CHAPTER 9. RESISTANCE AND EQUIVARIANCE 280
and that γ is the proportion of outliers. Then the mean vectors of the clusters
can be chosen to make the outliers bad leverage points. (This type of data set
is frequently used in simulations where the affine and regression equivariance
of the estimators is used to justify these choices.) It is well known that the
LMS(cn ), LTA(cn ) and LTA(cn ) are defined by the “narrowest strip” cover-
ing cn of the cases where the width of the strip is measured in the vertical
direction with the L∞ , L1 , and L2 criterion, respectively. We will assume
that cn ≈ n/2 and focus on the LMS estimator since the narrowness of the
strip is simply the vertical width of the strip.
Figure 9.1 will be useful for examining the resistance of the LMS estima-
tor. The data set consists of 300 N2(0, I 2) clean cases and 200
cases. Then the narrowest strip that covers only clean cases covers 1/[2(1−γ)]
of the clean cases. For the artificial data, γ = 0.4, and 5/6 of the clean cases
are covered and the width of the strip is approximately 2.76. The strip
shown in Figure 9.1 consists of two parallel lines with y-intercepts of ±1.38
and covers approximately 250 cases. As this strip is rotated counterclockwise
about the origin until it is parallel to the y-axis, the vertical width of the
strip increases to ∞. Hence LMS will correctly produce a slope near zero
if no outliers are present. Next, stop the rotation when the center of the
strip passes through the center of both clusters, covering nearly 450 cases.
The vertical width of the strip can be decreased to a value less than 2.76
while still covering 250 cases. Hence the LMS fit will accommodate the
outliers, and with 40% contamination, an outlying cluster can tilt the LMS
fit considerably. As c → 0, the cluster of outliers tends to a point mass and
even greater tilting is possible; nevertheless, for the Figure 9.1 data, a 40%
point mass can not drive the LMS slope to ∞.
Next suppose that the 300 distinct clean cases lie exactly on the line
through the origin with zero slope. Then an “exact fit” to at least half of the
data occurs and any rotation from this line can cover at most 1 of the clean
cases. Hence a point mass will not be able to rotate LMS unless it consists
of at least 299 cases (creating 300 additional exact fits). Similar results hold
for the LTA and LTS estimators.
These remarks suggest that the narrowest band interpretation of the LTx
estimators gives a much fuller description of their resistance than their break-
down value. Also, setting β = 0 may lead to misleading simulation studies.
CHAPTER 9. RESISTANCE AND EQUIVARIANCE 282
10
8
6
4
Y
2
0
-2
-4
-2 0 2 4 6 8 10
Figure 9.1: 300 N(0, I 2 ) cases and 200 N((9, 9)T , 0.25I 2 ) cases
9.5 Complements
Feller (1957) is a great source for examining subsampling behavior when the
data set is fixed.
Hampel, Ronchetti, Rousseeuw and Stahel (1986, p. 96-98) and Donoho
and Huber (1983) provide some history for breakdown. Maguluri and Singh
(1997) have interesting examples on breakdown. Morgenthaler (1989) and
Stefanski (1991) conjectured that high breakdown estimators with high effi-
ciency are not possible. These conjectures have been shown to be false.
9.6 Problems
9.1 a) Enter or download the following R/Splus function
pifclean<-
function(k, gam)
{
p <- floor(log(3/k)/log(1 - gam))
list(p = p)
}
b) Include the output from the commands below that are used to produce
the second column of Table 9.1.
Multivariate Models
284
CHAPTER 10. MULTIVARIATE MODELS 285
There are some differences in the notation used in multiple linear regres-
sion and multivariate location dispersion models. Notice that W could be
used as the design matrix in multiple linear regression although usually the
first column of the regression design matrix is a vector of ones. The n × p de-
sign matrix in the multiple linear regression model was denoted by X and Xi
denoted the ith column of X. In the multivariate location dispersion model,
X and X i will be used to denote a p × 1 random vector with observed value
xi , and xTi is the ith row of the data matrix W . Johnson and Wichern (1988,
p. 7, 53) uses X to denote the n × p data matrix and a n × 1 random vector,
relying on the context to indicate whether X is a random vector or data
matrix. Software tends to use different notation. For example, R/Splus will
use commands such as
var(x)
and
E(AX) = AE(X) and E(AXB) = AE(X)B. (10.3)
Thus
Cov(a + AX) = Cov(AX) = ACov(X)AT . (10.4)
Cov(X) = Σ.
Proposition 10.2. a) All subsets of a MVN are MVN: (Xk1 , ..., Xkq )T
∼ Nq (µ̃, Σ̃) where µ̃i = E(Xki ) and Σ̃ij = Cov(Xki , Xkj ). In particular,
X 1 ∼ Nq (µ1 , Σ11 ) and X 2 ∼ Np−q (µ2, Σ22).
b) If X 1 and X 2 are independent, then Cov(X 1 , X 2 ) = Σ12 =
E[(X 1 − E(X 1 ))(X 2 − E(X 2 ))T ] = 0, a q × (p − q) matrix of zeroes.
c) If X ∼ Np (µ, Σ), then X 1 and X 2 are independent iff Σ12 = 0.
d) If X 1 ∼ Nq (µ1, Σ11) and X 2 ∼ Np−q (µ2 , Σ22) are independent, then
X1 µ1 Σ11 0
∼ Np , .
X2 µ2 0 Σ22
Example 10.1. Let p = 2 and let (Y, X)T have a bivariate normal
distribution. That is,
Y µY σY2 Cov(Y, X)
∼ N2 , 2 .
X µX Cov(X, Y ) σX
Also recall that the population correlation between X and Y is given by
Cov(X, Y ) σX,Y
ρ(X, Y ) = =
VAR(X) VAR(Y ) σX σY
if σX > 0 and σY > 0. Then Y |X = x ∼ N(E(Y |X = x), VAR(Y |X = x))
where the conditional mean
1 σY2
E(Y |X = x) = µY + Cov(Y, X) 2 (x − µX ) = µY + ρ(X, Y ) 2
(x − µX )
σX σX
CHAPTER 10. MULTIVARIATE MODELS 288
1 1 −1 2 2 1 1
exp( (x + 2ρxy + y )) ≡ f1 (x, y) + f2(x, y)
2 2π 1 − ρ2 2(1 − ρ2 ) 2 2
where x and y are real and 0 < ρ < 1. Since both marginal distributions of
fi (x, y) are N(0,1) for i = 1 and 2 by Proposition
10.2 a), the marginal dis-
tributions of X and Y are N(0,1). Since xyfi (x, y)dxdy = ρ for i = 1 and
−ρ for i = 2, X and Y are uncorrelated, but X and Y are not independent
since f(x, y) = fX (x)fY (y).
Remark 10.2. In Proposition 10.3, suppose that X = (Y, X2 , ..., Xp)T .
Let X1 = Y and X 2 = (X2 , ..., Xp)T . Then E[Y |X 2 ] = β1 +β2X2 +· · ·+βp Xp
and VAR[Y |X 2] is a constant that does not depend on X 2 . Hence Y =
β1 + β2X2 + · · · + βpXp + e follows the multiple linear regression model.
CHAPTER 10. MULTIVARIATE MODELS 289
E(X) = µ (10.7)
and
Cov(X) = cX Σ (10.8)
where
cX = −2ψ (0).
has density
π p/2
h(u) = kp up/2−1g(u). (10.10)
Γ(p/2)
E(X|B T X) = µ + M B B T (X − µ) = aB + M B B T X (10.11)
aB = µ − M B B T µ = (I p − M B B T )µ,
and
M B = ΣB(B T ΣB)−1 .
Notice that in the formula for M B , Σ can be replaced by cΣ where c > 0 is
a constant. In particular, if the EC distribution has 2nd moments, Cov(X)
can be used instead of Σ.
Proposition 10.5. Let X ∼ ECp (µ, Σ, g) and assume that E(X) exists.
b) Even if the first moment does not exist, the conditional median
MED(Y |X) = α + βT X
Now
Y Y Y
E[ | X] = E[ |B T
]
X X X
−1 Y − µY
= µ + ΣB(B ΣB) B T T
X − µX
by Lemma 10.4. The right hand side of the last equation is equal to
ΣY X −1 µY − ΣY X Σ−1 −1
XX µX + ΣY X ΣXX X
µ+ ΣXX (X − µX ) =
ΣXX X
βT = ΣY X Σ−1
XX .
b) See Croux, Dehon, Rousseeuw and Van Aelst (2001) for references.
Example 10.2. This example illustrates another application of Lemma
10.4. Suppose that X comes from a mixture of two multivariate normals
with the same mean and proportional covariance matrices. That is, let
where c > 0 and 0 < γ < 1. Since the multivariate normal distribution is
elliptically contoured (and see Proposition 4.1c),
and
1
n
C(W ) = S = (xi − T (W ))(xi − T (W ))T
n − 1 i=1
CHAPTER 10. MULTIVARIATE MODELS 294
and will be denoted by MDi2 . When T (W ) and C(W ) are estimators other
than the sample mean and covariance, Di2 will sometimes be denoted by
RDi2 .
The following proposition shows that the Mahalanobis distances are in-
variant under affine transformations. See Rousseeuw and Leroy (1987, p.
252-262) for similar results.
Proposition 10.7. If (T, C) is affine equivariant, then
Di2 (W ) ≡ Di2 (T (W ), C(W )) =
Di2 (T (Z), C(Z)) ≡ Di2 (Z). (10.16)
10.5 Breakdown
This section gives a standard definition of breakdown (see Zuo 2001 for refer-
ences) for estimators of multivariate location and dispersion. The following
notation will be useful. Let W denote the data matrix where the ith row
corresponds to the ith case. For multivariate location and dispersion, W is
the n × p matrix with ith row xTi . Let W nd denote the data matrix with
ith row wTi where any d of the cases have been replaced by arbitrarily bad
contaminated cases. Then the contamination fraction is γ = d/n.
Let (T (W ), C(W )) denote an estimator of multivariate location and dis-
persion where the p× 1 vector T (W ) is an estimator of location and the p× p
symmetric positive semidefinite matrix C(W ) is an estimator of dispersion.
The breakdown value of the multivariate location estimator T at W is
d
B(T, W ) = min{ : sup T (W nd ) = ∞}
n Wn
d
1
cn
ci,j = (zi,k − z k )(zj,k − z j ).
cn − 1 k=1
Hence the maximum eigenvalue λ1 can not get arbitrarily large if the zi are
all contained in some ball of radius R about the origin, eg, if none of the
cn cases is an outlier. If all of the zi are bounded, then λ1 is bounded,
and λp can only be driven to zero if the determinant of C can be driven to
zero. The determinant |S| of S is known as the generalized sample variance.
Consider the hyperellipsoid
{z : (z − T )T C −1 (z − T ) ≤ D(c
2
n)
} (10.17)
2
where D(c n)
is the cn th smallest squared Mahalanobis distance based on
(T, C). This ellipsoid contains the cn cases with the smallest Di2 . The vol-
ume of this ellipsoid is proportional to the square root of the determinant
|C|1/2, and this volume will be positive unless extreme degeneracy is present
among the cn cases. See Johnson and Wichern (1988, p. 103-104).
a consistent estimator for (µ, aΣ) where a is some positive constant when the
data X i are elliptically contoured ECp (µ, Σ, g), and TM CD has a Gaussian
limit. See Butler, Davies, and Jhun (1993).
Computing robust covariance estimators can be very expensive. For ex-
ample, to compute the exact MCD(cn ) estimator (TM CD , CM CD ), we need to
consider the C(n, cn ) subsets of size cn . Woodruff and Rocke (1994, p. 893)
note that if 1 billion subsets of size 101 could be evaluated per second, it
would require 1033 millenia to search through all C(200, 101) subsets if the
sample size n = 200.
Hence high breakdown (HB) algorithms will again be used to approximate
the robust estimators. Many of the properties and techniques used for HB
regression algorithm estimators carry over for HB algorithm estimators of
multivariate location and dispersion. Elemental sets are the key ingredient
for both basic resampling and concentration algorithms.
Definition 10.9. Suppose that x1 , ..., xn are p × 1 vectors of observed
data. An elemental set J is a set of p + 1 cases in the multivariate location
and dispersion model. An elemental start is the sample mean and sample co-
variance matrix of the data corresponding to J. In a concentration algorithm,
let (T0,j , C 0,j ) be the jth start (not necessarily elemental) and compute all n
Mahalanobis distances Di (T0,j , C 0,j ). At the next iteration, the classical es-
timator (T1,j , C 1,j ) = (x1,j , S 1,j ) is computed from the cn ≈ n/2 cases corre-
sponding to the smallest distances. This iteration can be continued for k steps
resulting in the sequence of estimators (T0,j , C 0,j ), (T1,j , C 1,j ), ..., (Tk,j , C k,j ).
The result of the iteration (Tk,j , C k,j ) is called the jth attractor. Kn starts
are used, and the concentration estimator, called the CMCD estimator, is the
attractor that has the smallest determinant det(C k,j ). The basic resampling
algorithm estimator is a special case where k = 0 so that the attractor is the
start: (xk,j , S k,j ) = (x0,j , S 0,j ).
This concentration algorithm is a simplified version of the algorithms
given by Rousseeuw and Van Driessen (1999) and Hawkins and Olive (1999a).
Using k = 10 concentration steps often works well.
Proposition 10.8: Rousseeuw and Van Driessen (1999, p. 214).
Suppose that the classical estimator (xi,j , S i,j ) is computed from cn cases and
that the n Mahalanobis distances RDk ≡ RDk (xi,j , S i,j ) are computed. If
(xi+1,j , S i+1,j ) is the classical estimator computed from the cn cases with the
smallest Mahalanobis distances RDk , then the MCD criterion det(S i+1,j ) ≤
CHAPTER 10. MULTIVARIATE MODELS 298
1 n c
ci,j = (zi,k − z i )(zj,k − zj ).
cn − 1
k=1
Hence the maximum eigenvalue λ1 can not get arbitrarily large if the zi are
all contained in some ball of radius R about the origin, eg, if none of the
cn cases is an outlier. If all of the z i are bounded, then all of the λi are
bounded, and λp can only be driven to zero if the determinant of C can
be driven to zero. The determinant |S J | of S J is known as the generalized
sample variance.
CHAPTER 10. MULTIVARIATE MODELS 300
{z : (z − xJ )T S −1 2
J (z − xJ ) ≤ d }. (10.19)
2π p/2 p
d det(SJ ), (10.20)
pΓ(p/2)
and this volume will be positive unless extreme degeneracy is present among
the cn cases. See Johnson and Wichern (1988, p. 103-104). If d2 = D(c 2
n)
,
the cn th smallest squared Mahalanobis distance based on (xJ , S J ), then the
hyperellipsoid contains the cn cases with the smallest Di2 . Using the above
ideas suggests the following robust estimator.
Definition 10.11. Let the Mth start (T0,M , C 0,M ) = (x0,M , S 0,M ) be
the classical estimator applied after trimming the M% of cases furthest in
Euclidean distance from the coordinatewise median MED(W ) where M ∈
{0, 50} (or use, eg, M ∈ {0, 50, 60, 70, 80, 90, 95, 98}). Then concentra-
tion steps are performed resulting in the Mth attractor (Tk,M , C k,M ) =
(xk,M , S k,M ), and the M = 0 attractor is the DGK estimator. Let (TA , C A )
correspond to the attractor that has the smallest determinant. The median
ball algorithm (MBA) estimator (TM BA, C M BA) takes TM BA = TA and
MED(Di2 (TA , C A ))
C M BA = CA (10.21)
χ2p,0.5
has density
π p/2
h(u) = kp up/2−1g(u), (10.23)
Γ(p/2)
and the 50% highest density region has the form of the hyperellipsoid
{z : (z − µ)T Σ−1 (z − µ) ≤ U0.5 }
where U0.5 is the median of the distribution of U. For example, if the x are
MVN, then U has the χ2p distribution.
Remark 10.4. √
a) Butler, Davies and Jhun (1993): The MCD(cn ) estimator is a n con-
sistent HB estimator for (µ, aM CD Σ) where the constant aM CD > 0 depends
on the EC distribution.
b) Lopuhaä (1999): If (T, C) is a consistent estimator for (µ, aΣ) with
rate nδ where the constants a > 0 and δ > 0, then the classical estimator
(xM , S M ) computed after trimming the M% (where 0 < M < 100) of cases
with the largest distances Di (T, C) is a consistent estimator for (µ, aM Σ)
with the same rate nδ where aM > 0 is some constant. Notice that applying
the classical estimator to the cn ≈ n/2 cases with the smallest distances
corresponds to M = 50.
c) Rousseeuw and Van Driessen (1999): Assume that the classical esti-
mator (xm,j , S m,j ) is computed from cn cases and that the n Mahalanobis
distances Di ≡ Di (xm,j , S m,j ) are computed. If (xm+1,j , S m+1,j ) is the clas-
sical estimator computed from the cn cases with the smallest Mahalanobis
distances Di , then the MCD criterion det(S m+1,j ) ≤ det(S m,j ) with equality
iff (xm+1,j , S m+1,j ) = (xm,j , S m,j ).
d) Pratt (1959): Let K be a fixed positive integer and let the constant
a > 0. Suppose that (T1, C 1 ), ..., (TK , C K ) are K consistent estimators of
(µ, aΣ) each with the same rate nδ . If (TA , C A ) is an estimator obtained by
choosing one of the K estimators, then (TA , C A ) is a consistent estimator of
(µ, aΣ) with rate nδ .
e) Olive (2002): Suppose that (Ti, C i ) are consistent estimators for (µ, aiΣ)
where ai > 0 for i = 1, 2. Let Di,1 and Di,2 be the corresponding distances
and let R be the set of cases with distances Di (T1 , C 1) ≤ MED(Di (T1 , C 1)).
Let rn be the correlation between Di,1 and Di,2 for the cases in R. Then
rn → 1 in probability as n → ∞.
f) Olive (2004a): (x0,50, S 0,50) is a high breakdown estimator. If the data
distribution is EC but not spherically symmetric, then for m ≥ 0, S m,50 under
CHAPTER 10. MULTIVARIATE MODELS 302
estimates the major axis and over estimates the minor axis of the highest
density region. Concentration reduces but fails to eliminate this bias. Hence
the estimated highest density region based on the attractor is “shorter” in
the direction of the major axis and “fatter” in the direction of the minor axis
than estimated regions based on consistent
√ estimators. Arcones (1995) and
Kim (2000) showed that x0,50 is a HB n consistent estimator of µ.
The following remarks help explain why the MBA estimator is robust.
Using k = 5 concentration steps often works well. The scaling makes C M BA
a better estimate of Σ if the data is multivariate normal MVN. See Equa-
tions (11.2) and (11.4). The attractor (Tk,0, C k,0 ) that uses the classical
estimator (0% trimming) as a start is the DGK estimator and has good
statistical properties. By Remark 10.4f, the start (T0,50, C 0,50) that uses
50% trimming is a high breakdown estimator. Since only cases xi such that
xi − MED(x) ≤ MED(xi − MED(x)) are used, the largest eigenvalue of
C 0,50 is bounded if fewer than half of the cases are outliers.
The geometric behavior of the start (T0,M , C 0,M ) with M ≥ 0 is simple.
If the data xi are MVN (or EC) then the highest density regions of the
data are hyperellipsoids. The set of x closest to the coordinatewise median
in Euclidean distance is a hypersphere. For EC data the highest density
ellipsoid and hypersphere will have approximately the same center as the
hypersphere, and the hypersphere will be drawn towards the longest axis of
the hyperellipsoid. Hence too much data will be trimmed in that direction.
For example, if the data are MVN with Σ = diag(1, 2, ..., p) then C 0,M will
underestimate the largest variances and overestimate the smallest variances.
Taking k concentration steps can greatly reduce but not eliminate the bias
of C k,M if the data is EC, and the determinant |C k,M | < |C 0,M | unless the
attractor is equal to the start by Remark 10.4c. The attractor (Tk,50, C k,50 )
is not affine equivariant but is resistant to gross outliers in that they will
initially be given weight zero if they are further than the median Euclidean
distance from the coordinatewise median. Gnanadesikan and Kettenring
(1972, p. 94) suggest an estimator similar to the attractor (Tk,50, C k,50 ), also
see Croux and Van Aelst (2002).
Next, we will compare several concentration algorithms with theory and
simulation. Let the CMCD algorithm use k > 1 concentration steps where
the final estimator is the attractor that has the smallest determinant (the
MCD criterion). We recommend k = 10 for the DGK estimator and k = 5
CHAPTER 10. MULTIVARIATE MODELS 303
(see Rousseeuw and Leroy 1987, p. 259) which is proportional to the volume
of the hyperellipsoid
{z : (z − xk,M )T S −1 2
k,M (z − xk,M ) ≤ d } (10.26)
where d2 =√MED(Di2 (xk,M , S k,M )). The following two theorems show how
to produce n consistent robust estimators from starts that use O(n) cases.
The following theorem shows that the MBA estimator has good statistical
properties.
Theorem 10.14. Suppose (E1) holds.
a) If (TA , C A ) is the attractor that minimizes the volume criterion (10.25),
√
then (TA , C A ) is a HB n consistent estimator of (µ, aM CD Σ).
b)If√(TA , C A ) is the attractor that minimizes det(Sk,M ), then (TA , C A ) is
a√HB n consistent estimator of (µ, aM CD Σ). Hence the MBA estimator is
a n consistent HB estimator.
CHAPTER 10. MULTIVARIATE MODELS 307
one of the MBA attractors will minimize the criterion value and the result
follows. If (E1) holds and the distribution is not spherically symmetric, then
the probability that the DGK attractor minimizes the determinant goes to
one as n → ∞, and (TCM CD , C CM CD ) is asymptotically equivalent to the
DGK estimator (Tk,0, C k,0 ). QED
To compare (TM BA, C M BA ) and (TF M CD , C F M CD ), we used simulated
data with n = 100 cases and computed the FMCD estimator with the
R/Splus function cov.mcd. Initially the data sets had no outliers, and all
100 cases were MVN with zero mean vector and Σ = diag(1,2, ..., p). We
generated 500 runs of this data with p = 4. The averaged diagonal elements
of C M BA were 1.202, 2.260, 3.237 and 4.204. (In the simulations, the scale
factor in Equation (10.21) appeared to be slightly too large for small n but
slowly converged to the correct factor as n increased.) The averaged diagonal
elements of C F M CD were 0.838, 1.697, 2.531, and 3.373. The approximation
1.2C F M CD ≈ Σ was good. For both matrices, all off diagonal elements had
average values less than 0.034 in magnitude.
Next data sets with 40% outliers were generated. The last 60 cases were
MVN with zero mean vector and Σ = diag(1,2, ..., p). The first √ 40 cases√were
MVN with the same Σ, but the p× 1 mean vector µ = (10, 10 2, ..., 10 p)T .
We generated 500 runs of this data using p = 4. Shown below are the averages
of C M BA and C F M CD . Notice that C F M CD performed extremely well while
the C M BA entries were over inflated by a factor of about 2 since the outliers
inflate the scale factor MED(Di2 (TA , C A ))/χ2p,0.5.
MBA FMCD
2.120 −0.031 −0.069 0.004 0.980 0.002 −0.004 0.011
−0.031 4.144 −0.111 −0.146 0.002 1.977 −0.008 −0.014
−0.069 −0.111 6.211 −0.419 −0.004 −0.008 2.991 0.013
−0.138 0.008 −0.419 7.933 0.011 −0.014 0.013 3.862
The DD plot of MDi versus RDi is useful for detecting outliers. The
resistant estimator will be useful if (T, C) ≈ (µ, cΣ) where c > 0 since
scaling by c affects the vertical labels of the RDi but not the shape of the
DD plot. For the outlier data, the MBA estimator is biased, but the outliers
in the MBA DD plot will have large RDi since C M BA ≈ 2C F M CD ≈ 2Σ.
When p is increased to 8, the cov.mcd estimator was usually not useful
for detecting the outliers for this type of contamination. Figure 10.1 shows
CHAPTER 10. MULTIVARIATE MODELS 309
FMCD DD Plot
72
4.0
34
3.5
19
3.0
RD
44
32
2.5
50
2.0
1.5
MD
Resistant DD Plot
30 6
20
27
20
15
RD
10
5
95 53
MD
that now the FMCD RDi are highly correlated with the MDi . The DD plot
based on the MBA estimator detects the outliers. See Figure 10.2.
10.8 Complements
The theory for concentration algorithms is due to Hawkins and Olive (2002)
and Olive and Hawkins (2006). The MBA estimator is due to Olive (2004a).
The computational and theoretical simplicity of the MBA estimator makes it
one of the most useful robust estimators ever proposed. An important appli-
cation of the robust algorithm estimators and of case diagnostics is to detect
outliers. Sometimes it can be assumed that the analysis for influential cases
and outliers was completely successful in classifying the cases into outliers
and good or “clean” cases. Then classical procedures can be performed on
the good cases. This assumption of perfect classification is often unreason-
able, and it is useful to have robust procedures, such as the MBA estimator,
that have rigorous asymptotic theory and are practical to compute. Since the
MBA estimator is about an order of magnitude faster than alternative robust
estimators, the MBA estimator may be useful for data mining applications.
In addition to concentration and randomly selecting elemental sets, three
other algorithm techniques are important. He and Wang (1996) suggest
computing the classical estimator and a robust estimator. The final cross
checking estimator is the classical estimator if both estimators are “close,”
otherwise the final estimator is the robust estimator. The second technique
was proposed by Gnanadesikan and Kettenring (1972, p. 90). They suggest
using the dispersion matrix C = [ci,j ] where ci,j is a robust estimator of the
covariance of Xi and Xj . Computing the classical estimator on a subset of
the data results in an estimator of this form. The identity
cases. Poor starts are discarded, and L of the best starts are evaluated on
the entire data set. This idea is also used by Rocke and Woodruff (1996) and
by Rousseeuw and Van Driessen (1999).
There certainly exist types of outlier configurations where the FMCD es-
timator outperforms the robust MBA estimator. The MBA estimator is most
vulnerable to outliers that lie inside the hypersphere based on the median
Euclidean distance from the coordinatewise median (see Problem 10.17 for a
remedy). Although the MBA estimator should not be viewed as a replace-
ment for the FMCD estimator, the FMCD estimator should be modified as
in Theorem 10.15. Until this modification appears in the software, both es-
timators can be used for outlier detection by making a scatterplot matrix of
the Mahalanobis distances from the FMCD, MBA and classical estimators.
The simplest version of the MBA estimator only has two starts. A simple
modification would be to add additional starts as in Problem 10.17.
Johnson and Wichern (1988) and Mardia, Kent and Bibby (1979) are
good references for multivariate statistical analysis based on the multivariate
normal distribution. The elliptically contoured distributions generalize the
multivariate normal distribution and are discussed (in increasing order of
difficulty) in Johnson (1987), Fang, Kotz, and Ng (1990), Fang and Anderson
(1990), and Gupta and Varga (1993). Fang, Kotz, and Ng (1990) sketch the
history of elliptically contoured distributions while Gupta and Varga (1993)
discuss matrix valued elliptically contoured distributions. Cambanis, Huang,
and Simons (1981), Chmielewski (1981) and Eaton (1986) are also important
references. Also see Muirhead (1982, p. 30–42).
Rousseeuw (1984) introduced the MCD and the minimum volume ellip-
soid MVE(cn) estimator. For the MVE estimator, T (W ) is the center of
the minimum volume ellipsoid covering cn of the observations and C(W )
is determined from the same ellipsoid. TM V E has a cube root rate and the
limiting distribution is not Gaussian. See Davies (1992). Bernholdt and
Fisher (2004) show that the MCD estimator can be computed with O(nv )
complexity where v = 1 + p(p + 3)/2 if x is a p × 1 vector.
Rocke and Woodruff (1996, p. 1050) claim that any affine equivariant
location and shape estimation method gives an unbiased location estimator
and a shape estimator that has an expectation that is a multiple of the true
shape for elliptically contoured distributions. Hence there are many can-
didate robust estimators of multivariate location and dispersion. See Cook,
Hawkins, and Weisberg (1993) for an exact algorithm for the MVE. Other pa-
pers on robust algorithms include Hawkins (1993b, 1994), Hawkins and Olive
CHAPTER 10. MULTIVARIATE MODELS 312
(1999a), Hawkins and Simonoff (1993), He and Wang (1996), Rousseeuw and
Van Driessen (1999), Rousseeuw and van Zomeren (1990), Ruppert * (1992),
and Woodruff and Rocke (1993). Rousseeuw and Leroy (1987, 7.1) also
describes many methods. Papers by Fung (1993), Ma and Genton (2001),
Olive (2004a), Poston, Wegman, Priebe, and Solka (1997) and Ruiz-Gazen
(1996) may also be of interest.
The discussion by Rocke and Woodruff (2001) and by Hubert (2001) of
Peña and Prieto (2001) stresses the fact that no one estimator can domi-
nate all others for every outlier configuration. These papers and Wisnowski,
Simpson, and Montgomery (2002) give outlier configurations that can cause
problems for the FMCD estimator.
10.9 Problems
10.1∗. Suppose that
X1 49 3 1 −1 0
X2 100 1 6 1 −1
∼ N4 , .
X3 17 −1 1 4 0
X4 7 0 −1 0 2
Σ−1 −1
XX ΣXY = [Cov(X)] Cov(X, Y ).
10.8. Using the notation under Lemma 10.4, show that if X is elliptically
contoured, then the conditional distribution of X 1 given that X 2 = x2 is
also elliptically contoured.
10.9∗. Suppose Y ∼ Nn (Xβ, σ 2I). Find the distribution of
(X T X)−1 X T Y if X is an n × p full rank constant matrix.
10.10. Recall that Cov(X, Y ) = E[(X − E(X))(Y − E(Y ))T ]. Using
the notation of Proposition 10.6 on p. 291, let (Y, X T )T be ECp+1 (µ, Σ, g)
where Y is a random variable. Let the covariance matrix of (Y, X T ) be
ΣY Y ΣY X VAR(Y ) Cov(Y, X)
Cov((Y, X ) ) = c
T T
=
ΣXY ΣXX Cov(X, Y ) Cov(X)
α = µY − β T µX and
β = [Cov(X)]−1Cov(X, Y ).
10.11. (Due to R.D. Cook.) Let X be a p × 1 random vector with
E(X) = 0 and Cov(X) = Σ. Let B be any constant full rank p × r matrix
where 1 ≤ r ≤ p. Suppose that for all such conforming matrices B,
E(X|B T X) = M B B T X
covmba, will display the code for the function. Use the args command, eg
args(covmba), to display the needed arguments for the function.
10.12. a) Download the maha function that creates the classical Maha-
lanobis distances.
b) Enter the following commands and check whether observations 201–210
look like outliers.
10.13. a) Download the rmaha function that creates the robust Maha-
lanobis distances.
b) Obtain outx2 as in Problem 10.12 b). R users need to enter the
command library(lqs). Enter the command rmaha(outx2) and check whether
observations 201–210 look like outliers.
10.14. a) Download the covmba function.
b) Download the program rcovsim.
c) Enter the command rcovsim(100) three times and include the output
in Word.
d) Explain what the output is showing.
10.15∗. a) Assuming that you have done the two source commands above
Problem 10.12 (and in R the library(lqs) command), type the command
ddcomp(buxx). This will make 4 DD plots based on the DGK, MBA, FMCD
and median ball estimators. The DGK and median ball estimators are the
two attractors used by the MBA estimator. With the leftmost mouse button,
move the curser to each outlier and click. This data is the Buxton (1920)
data and cases with numbers 61, 62, 63, 64, and 65 were the outliers with
head lengths near 5 feet. After identifying the outliers in each plot, hold the
rightmost mouse button down (and in R click on Stop) to advance to the
next plot. When done, hold down the Ctrl and c keys to make a copy of the
plot. Then paste the plot in Word.
CHAPTER 10. MULTIVARIATE MODELS 316
CMCD Applications
11.1 DD Plots
A basic way of designing a graphical display is to arrange for reference
situations to correspond to straight lines in the plot.
Chambers, Cleveland, Kleiner, and Tukey (1983, p. 322)
317
CHAPTER 11. CMCD APPLICATIONS 318
med(Di (A))/med(MDi )
CHAPTER 11. CMCD APPLICATIONS 319
which is generally not one. By taking τ = med(MDi )/med(Di (A)), the plot
will follow the identity line if (x, S) is a consistent estimator of (µ, cxΣ) and
if (TA , C A ) is a consistent estimator of (µ, aA Σ). (Using the notation from
Proposition 11.1, let (a1, a2) = (cx, aA ).) The classical estimator is consistent
if the population has second moments, and the algorithm estimator (TA , C A )
tends to be consistent on the class of EC distributions and biased otherwise.
By replacing the observed median med(MDi ) of the classical Mahalanobis
distances with the target population analog, say MED, τ can be chosen so
that the DD plot is simultaneously a diagnostic for elliptical symmetry and a
diagnostic for the target EC distribution. That is, the plotted points follow
the identity line if the data arise from a target EC distribution such as the
multivariate normal distribution, but the points follow a line with non-unit
slope if the data arise from an alternative EC distribution. In addition the
DD plot can often detect departures from elliptical symmetry such as outliers,
the presence of two groups, or the presence of a mixture distribution. These
facts make the DD plot a useful alternative to other graphical diagnostics for
target distributions. See Easton and McCulloch (1990), Li, Fang, and Zhu
(1997), and Liu, Parelius, and Singh (1999) for references.
Example 11.1. Rousseeuw and Van Driessen (1999) choose the multi-
variate normal Np (µ, Σ) distribution as the target. If the data are indeed iid
MVN vectors, then the (MDi )2 are asymptotically χ2p random variables, and
MED = χ2p,0.5 where χ2p,0.5 is the median of the χ2p distribution. Since the
target distribution is Gaussian, let
2
χp,0.5 χ2p,0.5
RDi = Di (A) so that τ = . (11.2)
med(Di (A)) med(Di (A))
Note that the DD plot can be tailored to follow the identity line if the
data are iid observations from any target elliptically contoured distribution
that has 2nd moments. If it is known that med(MDi ) ≈ MED where MED is
the target population analog (obtained, for example, via simulation, or from
the actual target distribution as in Equations (10.8), (10.9) and (10.10) on
p. 289), then use
MED
RDi = τ Di (A) = Di (A). (11.3)
med(Di (A))
CHAPTER 11. CMCD APPLICATIONS 320
the MBA starts as shown in Theorem 10.15. There exist data sets with out-
liers or two groups such that both the classical and robust estimators produce
ellipsoids that are nearly concentric. We suspect that the situation worsens
as p increases.
In a simulation study, Np (0, I p ) data were generated and cov.mcd was
used to compute first the Di (A), and then the RDi using Equation (11.2).
The results are shown in Table 11.1. Each choice of n and p used 100 runs,
and the 100 correlations between the RDi and the MDi were computed. The
mean and minimum of these correlations are reported along with the percent-
age of correlations that were less than 0.95 and 0.80. The simulation shows
that small data sets (of roughly size n < 8p + 20) yield plotted points that
may not cluster tightly about the identity line even if the data distribution
is Gaussian.
Since every estimator of location and dispersion defines an ellipsoid, the
DD plot can be used to examine which points are in the robust ellipsoid
{x : (x − TR )T C −1 2
R (x − TR ) ≤ RD(h) } (11.5)
2
where RD(h) is the hth smallest squared robust Mahalanobis distance, and
which points are in a classical ellipsoid
{x : (x − x)T S −1 (x − x) ≤ MD(h)
2
}. (11.6)
In the DD plot, points below RD(h) correspond to cases that are in the
ellipsoid given by Equation (11.5) while points to the left of MD(h) are in an
ellipsoid determined by Equation (11.6).
The DD plot will follow a line through the origin closely only if the two
ellipsoids are nearly concentric, eg if the data is EC. The DD plot will follow
the identity line closely if med(MDi ) ≈ MED, and RD2i =
MED
(xi − TA )T [( )2 C −1 T −1
A ](xi − TA ) ≈ (xi − x) S (xi − x) = MD2i
med(Di (A))
and (x, S) will often produce ellipsoids that are far from concentric.
CHAPTER 11. CMCD APPLICATIONS 322
12
10
3
8
RD
RD
6
2
4
1
2
0
0.5 1.0 1.5 2.0 2.5 3.0 3.5 0 1 2 3 4 5 6
MD MD
3.0
20
2.5
15
2.0
RD
RD
1.5
10
1.0
5
0.5
0
MD MD
omitting the cases with RDi ≥ χ2p,.975. This technique can magnify features
that are obscured when large RDi ’s are present. If the distribution of x is EC,
Proposition 11.1 implies that the correlation of the points in the weighted
DD plot will tend to one and that the points will cluster about a line passing
through the origin. For example, the plotted points in the weighted DD plot
(not shown) for the non-MVN EC data of Figure 11.1b are highly correlated
and still follow a line through the origin with a slope close to 2.0.
Figures 11.1c and 11.1d illustrate how to use the weighted DD plot. The
ith case in Figure 11.1c is (exp(xi,1), exp(xi,2), exp(xi,3))T where xi is the
ith case in Figure 11a; ie, the marginals follow a lognormal distribution.
The plot does not resemble the identity line, correctly suggesting that the
distribution of the data is not MVN; however, the correlation of the plotted
points israther high. Figure 11.1d is the weighted DD plot where cases with
RDi ≥ χ23,.975 ≈ 3.06 have been removed. Notice that the correlation of the
plotted points is not close to one and that the best fitting line in Figure 11.1d
may not pass through the origin. These results suggest that the distribution
of x is not EC.
It is easier to use the DD plot as a diagnostic for a target distribution
such as the MVN distribution than as a diagnostic for elliptical symmetry.
If the data arise from the target distribution, then the DD plot will tend
to be a useful diagnostic when the sample size n is such that the sample
correlation coefficient in the DD plot is at least 0.80 with high probability.
As a diagnostic for elliptical symmetry, it may be useful to add the OLS line
to the DD plot and weighted DD plot as a visual aid, along with numerical
quantities such as the OLS slope and the correlation of the plotted points.
Numerical methods for transforming data towards a target EC distribu-
tion have been developed. Generalizations of the Box–Cox transformation
towards a multivariate normal distribution are described in Velilla (1993).
Alternatively, Cook and Nachtsheim (1994) offer a two-step numerical pro-
cedure for transforming data towards a target EC distribution. The first step
simply gives zero weight to a fixed percentage of cases that have the largest
robust Mahalanobis distances, and the second step uses Monte Carlo case
reweighting with Voronoi weights.
Example 11.2. Buxton (1920, p. 232-5) gives 20 measurements of 88
CHAPTER 11. CMCD APPLICATIONS 324
64 61 63
300
65
62
200
RD
100
0
1 2 3 4 5
MD
2
1
1 2 3
MD
distribution and 40 iid cases from a bivariate normal distribution with mean
(0, −3)T and covariance I 2 . Figure 11.3 shows the classical covering ellipsoid
that uses (T, C) = (x, S). The symbol “1” denotes the data while the symbol
“2” is on the border of the covering ellipse. Notice that the classical ellipsoid
covers almost all of the data. Figure 11.4 displays the resistant covering
ellipse. The resistant covering ellipse contains most of the 100 “clean” cases
and excludes the 40 outliers. Problem 11.5 recreates similar figures with the
classical and the resistant R/Splus cov.mcd estimators.
Example 11.4. Buxton (1920) gives various measurements on 88 men
including height and nasal height. Five heights were recorded to be about
19mm and are massive outliers. Figure 11.5 shows that the classical covering
ellipsoid is quite large but does not include any of the outliers. Figure 11.6
shows that the resistant covering ellipsoid is not inflated by the outliers.
CHAPTER 11. CMCD APPLICATIONS 326
1
22
22 22
22
22 22
22 1 2222
2 1
2 1 1 2
1 1 1 22
22 2
2
22 11
11 2
22
22 1 11 1 2
2
2 1 1 111 1 2
2
222 1 11 1
2
2
2
2 1 1 1 1 22
2
22 1111 11 1111 11 11 2
2
2
2 1 1
2 11 1 1
22 11 11 1 22
1
1 11 22
0
2
222 1 111 2
22
22 1 1111 22
2
2 1 11111 1111 1
1
22
22 11 22
22 1 1 1
1 1 22
2 1 1 22
22 1 1 11 2
22 11 1 1 22
2 1 1 1 22
2
-2
2
2 1
1 1 1 1 22
2
2 1 1 11 1 2
2 11 1 1 1 2
2
2 1 22
2 1 11 1 2
2
21
2 1 1 1 1 2222
22 1 1
2 1 1 1 2
22
22 122221 1
1
2 1 22
1 222
-4
22 1 11
-3 -2 -1 0 1 2 3
1 2
2 22
2 22
22
2 22
2 1 22
2 22
22 1 1 11 2
1 1
22
2
2 1 11 22
2 1 1 1 1 22
22 1 11111 22
2 1 11 1 22
2 2
22 11 1 1 22
22
22 1 111111 11111 11 11 222
22 1
111 1 1 2
2 1 2
2 1 11 11 1 2
22
0
2
22 11 111 22
22 2
1 111 22
1 111111111 2 1
2 11 22 1
222 1 1 11
1 222
222 1 1 2
2 1
2 1 222 11
22 11 1 1 22
22 11 1
22 22
-2
2 1 1 2 1
22 11
22
2 11 111
22 11 1 1 1
22
2 2 1 1
2 22 1 1 1
22 1 22
22 2 1 1 1
222 22 1 1 11 1
22
22 22
2 1 1 1 1
1
22 22
2 1
-4
22 2 1 11 1
-4 -2 0 2 4
1
22 1 1
1 1 1 2
2
22 1 1 1 1 1
1 1 1 1
1 1 1 1 1 2 1
2
22 1 1 1 1 1
1 1
1 1 1
1 1 1 1 1 1
2
2
22 1 1 1
1 1
1 1 1 1 1
1 2
22 1
1
1
1 1
1 1 1
1
2
2
22
2 1 22
2 22
1500
2 22
2 22
2
2 22
2
2 2
2 2
2
2
22 22
2 2
22 2
22 2
22
22
22 22
1000
22 2
22
22
2 222
2
2 22
22
22 22
2 2
500
1 1 1 1 1
0
45 50 55 60
2 1 1 22
2 1 1 1
2 2 1 1 1 1 1 1 1 1 1 1 1 2
1222
2 2 1
1 1 1 1 1 1
1 1 1
22 1 1 1 1 1
1 1 1 1 1 1 22
22
2
1 1 1
1 1
1 1 1 1 1
1 22
22 1
1
1
1 1
1 1 1
1 2 2 2
2 2
2 21 2 2 2 2
1500
1000
500
1 1 1 1 1
0
40 45 50 55 60
{x : (x − T )T C −1 (x − T ) ≤ D(j)
2
}. (11.8)
The ith case (yi, xTi )T is trimmed if Di > D(j) . Then an estimator of β is
computed from the untrimmed cases. For example, if j ≈ 0.9n, then about
10% of the cases are trimmed, and OLS or L1 could be used on the untrimmed
cases.
Recall that a forward response plot is a plot of the fitted values Ŷi versus
the response Yi and is very useful for detecting outliers. If the MLR model
holds and the MLR estimator is good, then the plotted points will scatter
about the identity line that has unit slope and zero intercept. The identity
line is added to the plot as a visual aid, and the vertical deviations from the
identity line are equal to the residuals since Yi − Ŷi = ri .
The resistant trimmed views estimator combines ellipsoidal trimming and
the forward response plot. First compute (T, C), perhaps using the MBA
estimator or the R/Splus function cov.mcd. Trim the M% of the cases with
the largest Mahalanobis distances, and then compute the MLR estimator β̂M
from the untrimmed cases. Use M = 0, 10, 20, 30, 40, 50, 60, 70, 80, and
T
90 to generate ten forward response plots of the fitted values β̂ M xi versus yi
using all n cases. (Fewer plots are used for small data sets if β̂ M can not be
computed for large M.) These plots are called “trimmed views.”
Definition 11.2. The trimmed views (TV) estimator β̂ T ,n corresponds
to the trimmed view where the bulk of the plotted points follow the identity
line with smallest variance function, ignoring any outliers.
Example 11.4 (continued). For the Buxton (1920) data, height was
the response variable while an intercept, head length, nasal height, bigonal
breadth, and cephalic index were used as predictors in the multiple linear
regression model. Observation 9 was deleted since it had missing values.
Five individuals, cases 61–65, were reported to be about 0.75 inches tall
CHAPTER 11. CMCD APPLICATIONS 329
with head lengths well over five feet! OLS was used on the untrimmed cases
and Figure 11.7 shows four trimmed views corresponding to 90%, 70%, 40%
and 0% trimming. The OLS TV estimator used 70% trimming since this
trimmed view was best. Since the vertical distance from a plotted point
to the identity line is equal to the case’s residual, the outliers had massive
residuals for 90%, 70% and 40% trimming. Notice that the OLS trimmed
view with 0% trimming “passed through the outliers” since the cluster of
outliers is scattered about the identity line.
The TV estimator β̂ T ,n has good statistical properties if the estimator
applied to the untrimmed cases (X M,n , Y M,n ) has good statistical properties.
Candidates include OLS, L1 , Huber’s M–estimator, Mallows’ GM–estimator
or the Wilcoxon rank estimator. See Rousseeuw and Leroy (1987, p. 12-13,
150). The basic idea is that if an estimator with OP (n−1/2 ) convergence rate
is applied to a set of nM ∝ n cases, then the resulting estimator β̂ M,n also
has OP (n−1/2) rate provided that the response y was not used to select the
nM cases in the set. If β̂M,n − β = OP (n−1/2 ) for M = 0, ..., 90 then
β̂T ,n − β = OP (n−1/2) by Pratt (1959).
Let X n = X 0,n denote the full design matrix. Often when proving asymp-
totic normality of an MLR estimator β̂0,n , it is assumed that
X Tn X n
→ W −1 .
n
If β̂0,n has OP (n−1/2) rate and if for big enough n all of the diagonal elements
of −1
X TM,n X M,n
n
are all contained in an interval [0, B) for some B > 0, then β̂M,n − β =
OP (n−1/2).
The distribution of the estimator β̂M,n is especially simple when OLS is
used and the errors are iid N(0, σ 2 ). Then
β̂ M,n = (X TM,n X M,n )−1 X TM,n Y M,n ∼ Np (β, σ 2(X TM,n X M,n)−1 )
√
and n(β̂M,n −β) ∼ Np (0, σ 2(X TM,n X M,n/n)−1 ). Notice that this result does
not imply that the distribution of β̂ T ,n is normal.
CHAPTER 11. CMCD APPLICATIONS 330
90% 70%
1500
1500
1000
1000
y
y
500
500
0
0
2000 4000 6000 8000 10000 5000 10000 15000 20000
fit fit
40% 0%
1500
1500
1000
1000
y
y
500
500
0
fit fit
Table 11.2 compares the TV, MBA (for MLR), lmsreg, ltsreg, L1 and
OLS estimators on 7 data sets available from the text’s website. The column
headers give the file name while the remaining rows of the table give the
sample size n, the number of predictors p, the amount of trimming M used by
the TV estimator, the correlation of the residuals from the TV estimator with
the corresponding alternative estimator, and the cases that were outliers.
If the correlation was greater than 0.9, then the method was effective in
detecting the outliers, and the method failed, otherwise. Sometimes the
trimming percentage M for the TV estimator was picked after fitting the
bulk of the data in order to find the good leverage points and outliers.
Notice that the TV, MBA and OLS estimators were the same for the
Gladstone data and for the major data (Tremearne 1911) which had two
small y–outliers. For the Gladstone data, there is a cluster of infants that are
good leverage points, and we attempt to predict brain weight with the head
measurements height, length, breadth, size and cephalic index. Originally, the
variable length was incorrectly entered as 109 instead of 199 for case 119, and
the glado data contains this outlier. In 1997, lmsreg was not able to detect
the outlier while ltsreg did. Due to changes in the Splus 2000 code, lmsreg
CHAPTER 11. CMCD APPLICATIONS 331
Table 11.2: Summaries for Seven Data Sets, the Correlations of the Residuals
from TV(M) and the Alternative Method are Given in the 1st 5 Rows
√ Wi ,
Definition 11.3. For a multiple linear regression√model with weights
a weighted forward response plot is √ a plot of Wi xTi β̃ versus Wi Yi .
The weighted
√ residual
√ plot is a plot of Wi xTi β̃ versus the WMLR resid-
uals rW i = Wi Yi − Wi xi β̃.
T
Application 11.3. For resistant weighted MLR, use the WTV estimator
which is selected from ten weighted forward response plots.
CHAPTER 11. CMCD APPLICATIONS 332
11.5 Complements
The first section of this chapter followed Olive (2002) closely. The DD plot
can be used to diagnose elliptical symmetry, to detect outliers, and to assess
the success of numerical methods for transforming data towards an ellipti-
cally contoured distribution. Since many statistical methods assume that
the underlying data distribution is Gaussian or EC, there is an enormous
literature on numerical tests for elliptical symmetry. Bogdan (1999), Czörgö
(1986) and Thode (2002) provide references for tests for multivariate normal-
ity while Koltchinskii and Li (1998) and Manzotti, Pérez and Quiroz (2002)
have references for tests for elliptically contoured distributions.
The TV estimator was proposed by Olive (2002, 2005) and is similar to
an estimator proposed by Rousseeuw and van Zomeren (1992). Although
both the TV and MBA estimators have the good OP (n−1/2 ) convergence
rate, their efficiency under normality may be very low. (We could argue that
the TV and OLS estimators are asymptotically equivalent on clean data if
0% trimming is always picked when all 10 plots look good.) Using the TV
and MBA estimators as the initial estimator in the cross checking estimator
results in a resistant (easily computed but zero breakdown) asymptotically
efficient final estimator. High breakdown estimators that have high efficiency
tend to be impractical to compute, but an exception is the cross checking
estimator that uses the CLTS estimator from Theorem 8.7 as the initial
robust estimator.
The ideas used in Section 11.3 have the potential for making many meth-
ods resistant. First, suppose that the MLR model holds but Cov(e) = σ 2Σ
and Σ = V V where V is known and nonsingular. Then V −1 Y = V −1 Xβ+
V −1 e, and the TV and MBA MLR estimators can be applied to Ỹ = V −1 Y
and X̃ = V −1 X provided that OLS is fit without an intercept.
Secondly, many 1D regression models (where yi is independent of xi given
the sufficient predictor xTi β) can be made resistant by making EY plots
of the estimated sufficient predictor xTi β̂M versus yi for the 10 trimming
CHAPTER 11. CMCD APPLICATIONS 334
11.6 Problems
PROBLEMS WITH AN ASTERISK * ARE ESPECIALLY USE-
FUL.
11.1∗. If X and Y are random variables, show that
Cov(X, Y) = [Var(X + Y) − Var(X − Y)]/4.
R/Splus Problems
Warning: Use the command source(“A:/rpack.txt”) to download
the programs. See Preface or Section 14.2. Typing the name of the
rpack function, eg ddplot, will display the code for the function. Use the
args command, eg args(ddplot), to display the needed arguments for the
function.
11.2. a) Download the program ddsim. (In R, type the command li-
brary(lqs).)
b) Using the function ddsim for p = 2, 3, 4, determine how large the
sample size n should be in order for the DD plot of n Np(0, I p) cases to
be cluster tightly about the identity line with high probability. Table your
results. (Hint: type the command ddsim(n=20,p=2) and increase n by 10
until most of the 20 plots look linear. Then repeat for p = 3 with the n that
worked for p = 2. Then repeat for p = 4 with the n that worked for p = 3.)
11.3. a) Download the program corrsim. (In R, type the command
library(lqs).)
b) A numerical quantity of interest is the correlation between the MDi
and RDi in a DD plot that uses n Np (0, I p ) cases. Using the function corrsim
CHAPTER 11. CMCD APPLICATIONS 335
for p = 2, 3, 4, determine how large the sample size n should be in order for
9 out of 10 correlations to be greater than 0.9. (Try to make n small.) Table
your results. (Hint: type the command corrsim(n=20,p=2,nruns=10) and
increase n by 10 until 9 or 10 of the correlations are greater than 0.9. Then
repeat for p = 3 with the n that worked for p = 2. Then repeat for p = 4
with the n that worked for p = 3.)
11.4∗. a) Download the ddplot function. (In R, type the command
library(lqs).)
b) Using the following commands to make generate data from the EC
distribution (1 −
)Np (0, I p ) +
Np(0, 25 I p ) where p = 3 and
= 0.4.
n <- 400
p <- 3
eps <- 0.4
x <- matrix(rnorm(n * p), ncol = p, nrow = n)
zu <- runif(n)
x[zu < eps,] <- x[zu < eps,]*5
c) Use the command ddplot(x) to make a DD plot and include the plot
in Word. What is the slope of the line followed by the plotted points?
11.5. a) Download the ellipse function.
b) Use the following commands to create a bivariate data set with outliers
and to obtain a classical and robust covering ellipsoid. Include the two plots
in Word. (In R, type the command library(lqs).)
c) Use the command mplot(x) and place the resulting plot in Word.
d) Do you prefer the DD plot or the mplot? Explain.
11.7 a) Download the function wddplot.
b) Enter the commands in Problem 11.4b to obtain a data set x.
c) Use the command wddplot(x) and place the resulting plot in Word.
11.8. a) In addition to the source(“A:/rpack.txt”) command, also use
the source(“A:/robdata.txt”) command (and in R, type the library(lqs) com-
mand).
b) Type the command tvreg(buxx,buxy). Click the rightmost mouse but-
ton (and in R, highlight Stop). The forward response plot should appear.
Repeat 10 times and remember which plot percentage M (say M = 0) had
the best forward response plot. Then type the command tvreg2(buxx,buxy, M
= 0) (except use your value of M, not 0). Again, click the rightmost mouse
button (and in R, highlight Stop). The forward response plot should appear.
Hold down the Ctrl and c keys to make a copy of the plot. Then paste the
plot in Word.
c) The estimated coefficients β̂ T V from the best plot should have appeared
on the screen. Copy and paste these coefficients into Word.
Chapter 12
1D Regression
... estimates of the linear regression coefficients are relevant to the linear
parameters of a broader class of models than might have been suspected.
Brillinger (1977, p. 509)
After computing β̂, one may go on to prepare a scatter plot of the points
(β̂xj , yj ), j = 1, ..., n and look for a functional form for g(·).
Brillinger (1983, p. 98)
y x|βT x. (12.1)
337
CHAPTER 12. 1D REGRESSION 338
y = m(α + β T x) + e. (12.3)
t(y) = α + βT x + e.
Koenker and Geling (2001) note that if yi is an observed survival time, then
many survival models have the form of Equation (12.4). They provide three
illustrations including the Cox (1972) proportional hazards model. Li and
Duan (1989, p. 1014) note that the class of 1D regression models also includes
binary regression models, censored regression models, and certain projection
pursuit models. Applications are also given by Stoker (1986), Horowitz (1996,
1998) and Cook and Weisberg (1999a).
Definition 12.2. Regression is the study of the conditional distribution
of y|x. Focus is often on the mean function E(y|x) and/or the variance
function VAR(y|x). There is a distribution for each value of x = xo such
that y|x = xo is defined. For a 1D regression,
and
where M is the kernel mean function and V is the kernel variance function.
Notice that the mean and variance functions depend on the same linear
combination if the 1D regression model is valid. This dependence is typical of
GLM’s where M and V are known kernel mean and variance functions that
depend on the family of GLM’s. See Cook and Weisberg (1999a, section
23.1). A heteroscedastic regression model
y = M(β T1 x) + V (β T2 x) e (12.5)
βT x with “no loss of information.” Cook and Weisberg (1999a, p. 411) de-
fine a sufficient summary plot (SSP) to be a plot that contains all the sample
regression information about the conditional distribution y|x of the response
given the predictors.
Definition 12.4: If the 1D regression model holds, then y x|a + cβT x
for any constants a and c = 0. The quantity a + cβ T x is called a sufficient
predictor (SP), and a sufficient summary plot is a plot of any SP versus y. An
T
estimated sufficient predictor (ESP) is α̃+ β̃ x where β̃ is an estimator of cβ
for some nonzero constant c. An estimated sufficient summary plot (ESSP)
or EY plot is a plot of any ESP versus y.
If there is only one predictor x, then the plot of x versus y is both a
sufficient summary plot and an EY plot, but generally only an EY plot can
be made. Since a can be any constant, a = 0 is often used. The following
section shows how to use the OLS regression of y on x to obtain an ESP.
If the predictors x satisfy this condition, then for any given predictor xj ,
E[xj |βT x] = aj + tj β T x.
Notice that β is a fixed (p − 1) × 1 vector. If x is elliptically contoured (EC)
with 1st moments, then the assumption of linearly related predictors holds
since
E[x|bT x] = ab + tb bT x
for any nonzero (p − 1) × 1 vector b (see Lemma 10.4 on p. 290). The
condition of linearly related predictors is impossible to check since β is un-
known, but the condition is far weaker than the assumption that x is EC.
CHAPTER 12. 1D REGRESSION 341
The stronger EC condition is often used since there are checks for whether
this condition is plausible, eg use the DD plot. The following proposition
gives an equivalent definition of linearly related predictors. Both definitions
are frequently used in the regression graphics literature.
Proposition 12.1. The predictors x are linearly related iff
bTp x ap + tp β T x
QED
Following Cook (1998a, p. 143-144), assume that there is an objective
function
1
n
Ln (a, b) = L(a + bT xi , yi) (12.8)
n i=1
where L(u, v) is a bivariate function that is a convex function of the first
argument u. Assume that the estimate (â, b̂) of (a, b) satisfies
L(a + bT x, y) = (y − a − bT x)2 .
CHAPTER 12. 1D REGRESSION 342
500
Y
0
-500
-10 -5 0 5 10
SP
0
-500
-10 -5 0 5 10
- SP
usual multiple linear regression of yi on xi , but we are not assuming that the
multiple linear regression model holds; however, we are hoping that the 1D
regression model y x|βT x is a useful approximation to the data and that
b̂OLS ≈ cβ for some nonzero constant c. In addition to Theorem 12.2, nice
results exist if the single index model is appropriate. Recall that
Cov(x, y) = E[(x − E(x))((y − E(y))T ].
Definition 12.7. Suppose that (yi , xTi )T are iid observations and that
the positive definite (p − 1) × (p − 1) matrix Cov(x) = ΣX and the (p − 1) × 1
vector Cov(x, y) = ΣX,Y . Let the OLS estimator (â, b̂) be computed from
the multiple linear regression of y on x plus a constant. Then (â, b̂) estimates
the population quantity (αOLS , β OLS ) where
β OLS = Σ−1
X ΣX,Y . (12.10)
The following notation will be useful for studying the OLS estimator.
Let the sufficient predictor z = β T x and let w = x − E(x). Let r =
w − (ΣX β)β T w.
Theorem 12.3. In addition to the conditions of Definition 12.7, also
assume that yi = m(βT xi ) + ei where the zero mean constant variance iid
errors ei are independent of the predictors xi . Then
β OLS = Σ−1
X ΣX,Y = cm,X β + um,X (12.11)
where the scalar
cm,X = E[βT (x − E(x)) m(βT x)] (12.12)
and the bias vector
um,X = Σ−1
X E[m(β x)r].
T
(12.13)
Moreover, um,X = 0 if x is from an elliptically contoured distribution with
2nd moments, and cm,X = 0 unless Cov(x, y) = 0. If the multiple linear
regression model holds, then cm,X = 1, and um,X = 0.
The proof of the above result is outlined in Problem 12.2 using an argu-
ment due to Aldrin, Bφlviken, and Schweder (1993). See related results in
Stoker (1986) and Cook, Hawkins, and Weisberg (1992). If the 1D regres-
sion model is appropriate, then typically Cov(x, y) = 0 unless β T x follows a
symmetric distribution and m is symmetric about the median of β T x.
CHAPTER 12. 1D REGRESSION 346
Definition 12.8. Let (â, b̂) denote the OLS estimate obtained from the
OLS multiple linear regression of y on x. The OLS view is a plot of
T
b̂ x versus y.
Remark 12.4. All of this awkward notation and theory leads to one of
the most remarkable results in statistics, perhaps first noted by Brillinger
(1977, 1983) and called the 1D Estimation Result by Cook and Weisberg
(1999a, p. 432). The result is that if the 1D regression model is appropriate,
then the OLS view will frequently be a useful estimated sufficient summary
T
plot (ESSP). Hence the OLS predictor b̂ x is a useful estimated sufficient
predictor (ESP).
Although the OLS view is frequently a good ESSP if no strong nonlinear-
ities are present in the predictors and if cm,X = 0 (eg the sufficient summary
plot of βT x versus y is not approximately symmetric), even better estimated
sufficient summary plots can be obtained by using ellipsoidal trimming. This
topic is discussed in the following section and follows Olive (2002) closely.
for each vector of observed predictors xi . If the ordered distances D(j) are
CHAPTER 12. 1D REGRESSION 347
The two ideas of using ellipsoidal trimming to reduce the bias and choos-
ing a view with a smooth mean function and smallest variance function can
be combined into a graphical method for finding the estimated sufficient sum-
mary plot and the estimated sufficient predictor. Trim the M% of the cases
with the largest Mahalanobis distances, and then compute the OLS estima-
tor (α̂M , β̂ M ) from the untrimmed cases. Use M = 0, 10, 20, 30, 40, 50, 60,
T
70, 80, and 90 to generate ten plots of β̂ M x versus y using all n cases. In
analogy with the Cook and Weisberg procedure for visualizing 1D structure
with two predictors, the plots will be called “trimmed views.” Notice that
M = 0 corresponds to the OLS view.
Definition 12.9. The best trimmed view is the trimmed view with a
smooth mean function and the smallest variance function and is the estimated
sufficient summary plot. If M ∗ = E is the percentage of cases trimmed that
T
corresponds to the best trimmed view, then β̂ E x is the estimated sufficient
predictor.
The following examples illustrate the R/Splus function trviews that is
used to produce the ESSP. If R is used instead of Splus, the command
library(lqs)
trviews(X, Y)
Intercept X1 X2 X3
0.6701255 3.133926 4.031048 7.593501
Intercept X1 X2 X3
1.101398 8.873677 12.99655 18.29054
Intercept X1 X2 X3
0.9702788 10.71646 15.40126 23.35055
Intercept X1 X2 X3
0.5937255 13.44889 23.47785 32.74164
Intercept X1 X2 X3
1.086138 12.60514 25.06613 37.25504
Intercept X1 X2 X3
4.621724 19.54774 34.87627 48.79709
Intercept X1 X2 X3
3.165427 22.85721 36.09381 53.15153
Intercept X1 X2 X3
5.829141 31.63738 56.56191 82.94031
Intercept X1 X2 X3
4.241797 36.24316 70.94507 105.3816
Intercept X1 X2 X3
6.485165 41.67623 87.39663 120.8251
The function generates 10 trimmed views. The first plot trims 90% of the
cases while the last plot does not trim any of the cases and is the OLS view.
To advance a plot, press the right button on the mouse (in R, highlight
stop rather than continue). After all of the trimmed views have been
generated, the output is presented. For example, the 5th line of numbers in
T
the output corresponds to α̂50 = 1.086138 and β̂50 where 50% trimming was
used. The second line of numbers corresponds to 80% trimming while the
T
last line corresponds to 0% trimming and gives the OLS estimate (α̂0 , β̂ 0 ) =
(â, b̂). The trimmed views with 50% and 90% trimming were very good.
We decided that the view with 50% trimming was the best. Hence β̂E =
(12.60514, 25.06613, 37.25504)T ≈ 12.5β. The best view is shown in Figure
12.4 and is nearly identical to the sufficient summary plot shown in Figure
12.1. Notice that the OLS estimate = (41.68, 87.40, 120.83)T ≈ 42β. The
OLS view is Figure 1.6 in Chapter 1 (on p. 17) and is again very similar
to the sufficient summary plot, but it is not quite as smooth as the best
trimmed view.
The plot of the estimated sufficient predictor versus the sufficient predic-
CHAPTER 12. 1D REGRESSION 351
500
Y
0
-500
ESP
0
-5
-10
ESP
Figure 12.5: The angle between the SP and the ESP is nearly zero.
CHAPTER 12. 1D REGRESSION 352
tor is also informative. Of course this plot can usually only be generated for
simulated data since β is generally unknown. If the plotted points are highly
correlated (with |corr(ESP,SP)| > 0.95) and follow a line through the origin,
then the estimated sufficient summary plot is nearly as good as the sufficient
summary plot. The simulated data used β = (1, 2, 3)T , and the commands
0.6
0.2
0.2
Y
Y
−0.2
ESP ESP
0.6
0.2
0.2
Y
Y
−0.2
−0.2
ESP ESP
0.6
0.2
0.2
Y
Y
−0.2
−10 −8 −6 −4 −2 0 −0.2 0 2 4 6 8 10
ESP ESP
0.6
0.2
0.2
Y
Y
−0.2
−0.2
−12 −8 −6 −4 −2 0 1 2 3 4
ESP ESP
method b1 b2 b3
OLS View 0.0032 0.0011 0.0047
90% Trimmed OLS View 0.086 0.182 0.338
SIR View −0.394 −0.361 −0.845
10% Trimmed SIR VIEW −0.284 −0.473 −0.834
SAVE View −1.09 0.870 -0.480
40% Trimmed SAVE VIEW 0.256 0.591 0.765
PHD View −0.072 −0.029 −0.0097
90% Trimmed PHD VIEW −0.558 −0.499 −0.664
LMSREG VIEW −0.003 −0.005 −0.059
70% Trimmed LMSREG VIEW 0.143 0.287 0.428
not linear, and Figure 12.7 suggests that the mean function is neither linear
nor monotone.
Application 12.3. Assume that a known 1D regression model is as-
sumed for the data. Then the best trimmed view is a model checking plot
and can be used as a diagnostic for whether the assumed model is appropri-
ate.
The trimmed views are sometimes useful even when the assumption of
linearly related predictors fails. Cook and Li (2002) summarize when compet-
ing methods such as the OLS view, sliced inverse regression (SIR), principal
Hessian directions (PHD), and sliced average variance estimation (SAVE)
can fail. All four methods frequently perform well if there are no strong
nonlinearities present in the predictors.
Example 12.4 (continued). Figure 12.6 shows that the EY plots for
SIR, PHD, SAVE, and OLS are not very good while Figure 12.7 shows that
trimming improved the SIR, SAVE and OLS methods.
One goal for future research is to develop better methods for visualizing
1D regression. Trimmed views seem to become less effective as the number
of predictors k = p − 1 increases. Consider the sufficient predictor SP =
x1 + · · · + xk . With the sin(SP)/SP data, several trimming proportions gave
CHAPTER 12. 1D REGRESSION 356
0 1 2 3 4 5
0.6
0.4
Y
0.2
-0.2
5
4
ESP
3
2
1
0
40
30
SP
20
10
0
-0.2 0.0 0.2 0.4 0.6 0 10 20 30 40
good views with k = 3, but only one of the ten trimming proportions gave
a good view with k = 10. In addition to problems with dimension, it is
not clear which covariance estimator and which regression estimator should
be used. Preliminary investigations suggest that the classical covariance
estimator gives better estimates than cov.mcd, but among the many Splus
regression estimators, lmsreg often worked well. Theorem 12.2 suggests that
strictly convex regression estimators such as OLS will often work well, but
there is no theory for the robust regression estimators.
Example 12.4 continued. Replacing the OLS trimmed views by alter-
native MLR estimators often produced good EY plots, and for single index
models, the lmsreg estimator often worked the best. Figure 12.8 shows a
scatterplot matrix of y, ESP and SP where the sufficient predictor SP =
βT x. The ESP used ellipsoidal trimming with lmsreg instead of OLS. The
top row of Figure 12.8 shows that the estimated sufficient summary plot and
the sufficient summary plot are nearly identical. Also the correlation of the
CHAPTER 12. 1D REGRESSION 357
0.2
0.0
-0.2
0 1 2 3 4 5
FIT
ESP and the SP is nearly one. Table 12.1 shows the estimated sufficient pre-
dictor coefficients b when the sufficient predictor coefficients are c(1, 2, 3)T .
Only the SIR, SAVE, OLS and lmsreg trimmed views produce estimated
sufficient predictors that are highly correlated with the sufficient predictor.
Figure 12.9 helps illustrate why ellipsoidal trimming works. This view
used 70% trimming and the open circles denote cases that were trimmed.
The highlighted squares are the untrimmed cases. Note that the highlighted
cases are far more linear than the data set as a whole. Also lmsreg will give
half of the highlighted cases zero weight, further linearizing the function.
In Figure 12.9, the lmsreg constant α̂70 is included, and the plot is simply
the forward response plot of the weighted lmsreg fitted values versus y.
CHAPTER 12. 1D REGRESSION 358
The vertical deviations from the line through the origin are the “residuals”
T
yi − α̂70 − β̂70x and at least half of the highlighted cases have small residuals.
power transformations of Yeo and Johnson 2000.) Let λ be the power of the
transformation. Then the following four rules are often used.
The log rule states that positive predictors that have the ratio between
their largest and smallest values greater than ten should be transformed to
logs. See Cook and Weisberg (1999a, p. 87).
Secondly, if it is known that X2 ≈ X1λ and the ranges of X1 and X2 are
such that this relationship is one to one, then
1/λ
X1λ ≈ X2 and X2 ≈ X1 .
1/λ
Hence either the transformation X1λ or X2 will linearize the plot. This
relationship frequently occurs if there is a volume present. For example let
X2 be the volume of a sphere and let X1 be the circumference of a sphere.
Thirdly, the bulging rule states that changes to the power of X2 and the
power of X1 can be determined by the direction that the bulging side of the
curve points. If the curve is hollow up (the bulge points down), decrease the
power of X2 . If the curve is hollow down (the bulge points up), increase the
power of X2 If the curve bulges towards large values of X1 increase the power
of X1 . If the curve bulges towards small values of X1 decrease the power of
X1 . See Tukey (1977, p. 173–176).
Finally, Cook and Weisberg (1999a, p. 86) give the following rule.
To spread small values of a variable, make λ smaller.
To spread large values of a variable, make λ larger.
For example, in Figure 12.14c, small values of Y and large values of FFIT
need spreading, and using log(Y ) would make the plot more linear.
SP = α + βT x = α + βTS xS + β TE xE = α + β TS xS . (12.16)
The extraneous terms that can be eliminated given that the subset S is in
the model have zero coefficients.
Now suppose that I is a candidate subset of predictors, that S ⊆ I and
that O is the set of predictors not in I. Then
SP = α + β T x = α + β TS xS = α + βTS xS + β T(I/S)xI/S + 0T xO = α + β TI xI ,
(if I includes predictors from E, these will have zero coefficient). For any
subset I that includes all relevant predictors, the correlation
corr(α + βT xi, α + βT
I xI,i ) = 1. (12.17)
The difficulty with this approach is that fitting all of the possible sub-
models involves substantial computation. An exception to this difficulty is
multiple linear regression where there are efficient “leaps and bounds” algo-
rithms for searching all subsets when OLS is used (see Furnival and Wilson
1974). Since OLS often gives a useful ESP, the following all subsets procedure
can be used for 1D models.
• If the 1D ESP and the OLS ESP have ‘a strong linear relationship’ (for
example |corr(ESP, OLS ESP)| ≥ 0.95), then infer that the 1D problem
is one in which OLS may serve as an adequate surrogate for the correct
1D model fitting procedure.
• Perform a final check on the subsets that satisfy the Cp screen by using
them to fit the 1D model.
• The key to understanding which plots are the most useful is the obser-
vation that a wz plot is used to visualize the conditional distribution
of z given w. Since a 1D regression is the study of the conditional
distribution of y given α + β T x, the EY plot is used to visualize this
conditional distribution and should always be made. A major problem
with variable selection is that deleting important predictors can change
the functional form m of the model. In particular, if a multiple linear
CHAPTER 12. 1D REGRESSION 362
To see why the plots contain useful information, review Proposition 5.1
(on p. 140 - 142), and Remarks 5.2 and 5.3 with corr(r, rI ) replaced by
corr(V, V (I)). In many settings (not all of which meet the Li–Duan sufficient
conditions), the full model OLS ESP is a good estimator of the sufficient
predictor. If the fitted full 1D model y x|α+βT x is a useful approximation
to the data and if β̂ OLS is a good estimator of cβ where c = 0, then a
CHAPTER 12. 1D REGRESSION 363
subset I will produce an EY plot similar to the EY plot of the full model if
corr(OLS ESP, OLS ESP(I)) ≥ 0.95. Hence the EY plots based on the full
and submodel ESP can both be used to visualize the conditional distribution
of y.
To see that models with Cp(I) ≤ 2k for small k are interesting, assume
that subset I uses k predictors including the intercept, that Cp (I) ≤ 2k
and n ≥ 10p. Then 0.9 ≤ corr(V, V (I)), and both corr(V, V (I)) → 1.0 and
corr(OLS ESP, OLS ESP(I)) → 1.0 as n → ∞. Hence the plotted points in
both the VV plot and the EE plot will cluster about the identity line (see
Proposition 5.1 vi). Notice that for a fixed value of k, the model I with the
smallest value of Cp (I) maximizes corr(V, V (I)).
The Cp (I) ≤ k screen tends to overfit. We simulated multiple linear
regression and single index model data sets with p = 8 and n = 50, 100, 1000
and 10000. The true model S satisfied Cp (S) ≤ k for about 60% of the
simulated data sets, but S satisfied Cp (S) ≤ 2k for about 97% of the data
sets.
Assuming that a 1D model holds, a common assumption made for variable
selection is that the fitted full model ESP is a good estimator of the sufficient
predictor, and the usual numerical and graphical checks on this assumption
should be made. To see that this assumption is weaker than the assumption
that the OLS ESP is good, notice that if a 1D model holds but β̂OLS estimates
cβ where c = 0, then the Cp(I) criterion could wrongly suggest that all
subsets I have Cp (I) ≤ 2k. Hence we also need to check that c = 0.
There are several methods are for checking the OLS ESP, including: a) if
an ESP from an alternative fitting method is believed to be useful, check that
the ESP and the OLS ESP have a strong linear relationship – for example that
|corr(ESP, OLS ESP)| ≥ 0.95. b) Often examining the EY plot shows that a
1D model is reasonable. For example, if the data are tightly clustered about a
smooth curve, then a single index model may be appropriate. c) Verify that
a 1D model is appropriate using graphical techniques given by Cook and
Weisberg (1999a, p. 434-441). d) Verify that x has an elliptically contoured
distribution with 2nd moments and that the mean function m(α + β T x) is
not symmetric about the median of the distribution of α+β T x. Then results
from Li and Duan (1989) suggest that c = 0.
Condition a) is both the most useful (being a direct performance check)
and the easiest to check. A standard fitting method should be used when
available (eg, for parametric 1D models or the proportional hazards model).
CHAPTER 12. 1D REGRESSION 364
a) FF Plot b) RR Plot
40
15
30
10
20
FRES
FFIT
5
10
0
0
-5
-10
-10
-10 0 10 20 30 40 -5 0 5 10 15
SFIT SRES
50
40
40
30
30
Y
Y
20
20
10
10
0
-10 0 10 20 30 40 -10 0 10 20 30 40
FFIT SFIT
-0.6
-1.0
SRES
1
0
SFIT
2
2
1
FRES
0
0
Y
-1
-2
-2
-4
-4 -2 0 2 -4 -2 0 2
FFIT FFIT
2
2
1
SRES
0
Y
0
-2
-1
-4
-2
-4 -2 0 2 -4 -2 0 2
SFIT SFIT
Figure 12.12: Forward Response and Residual Plots for Boston Housing Data
20
2
15
0
RAD
Y
10
-2
5
-4
SFIT2 NOX
a) VV Plot b) EE Plot
15
60
10
40
FULLV
FESP
5
20
0
0
0 20 40 60 0 5 10 15
SUBV SESP
80
60
60
Y
Y
40
40
20
20
0
0 5 10 15 0 5 10 15
FESP SESP
three clusters of points in the plot of NOX versus RAD shown in Figure
12.13b (the single isolated point in the southeast corner of the plot actually
corresponds to several cases). The two clusters of high NOX and high RAD
points correspond to the cases with high per capita crime rate.
The tiny filled in triangles if Figure 12.13a represent the fitted values for
a quadratic. We added NOX 2 , RAD2 and NOX ∗ RAD to the full model
and again tried variable selection. Although the full quadratic in NOX and
RAD had a linear forward response plot, the submodel with NOX, RAD
and log(x2) was very similar. For this data set, NOX and RAD seem to be
the most important predictors, but other predictors are needed to make the
model linear and to reduce residual variation.
Example 12.8. In the Boston housing data, now let y = CRIM. Since
log(y) has a linear relationship with the predictors, y should follow a nonlin-
ear 1D regression model. Consider the full model with predictors log(x2), x3,
x4, x5 , log(x7 ), x8 , log(x9) and log(x12). Regardless of whether y or log(y)
is used as the response, the minimum Cp model from backward elimination
used a constant, log(x2 ), x4, log(x7 ), x8 and log(x12) as predictors. If y is
CHAPTER 12. 1D REGRESSION 369
the response, then the model is nonlinear and Cp = 5.699. Proposition 5.1
vi) (on p. 141) suggests that if Cp ≤ 2k, then the points in the VV plot
should tightly cluster about the identity line even if a multiple linear regres-
sion model fails to hold. Figure 12.14 shows the VV and EE plots for the
minimum Cp submodel. The EY plots for the full model and submodel are
also shown. Note that the clustering in the VV plot is indeed higher than
the clustering in the EE plot. Note that the EY plots are highly nonlinear
but are nearly identical.
12.5 Inference
Inference for tests of the form Ho: Aβ = 0 can be performed using χ2 tests
where A is a k × (p − 1) constant matrix of rank k. Let the 1D model
Y x|α + β T x be written as Y x|α + βTI xI + βTO xO where the reduced
model is Y x|α + βTI xI and xO denotes the terms outside of the reduced
model. Notice that test Ho: β = 0 and the test Ho: βO = 0 for a submodel
I have the correct form. The test for Ho: βi = 0 uses A = (0, ..., 0, 1, 0, ..., 0)
were the 1 is in the ith position. In the following theorem, it is crucial that
Ho: Aβ = 0. Tests for Ho: Aβ = 1, say, may not be valid even if the
sample size n is large. Let β̂ be the the OLS estimator of β and let the
−1 AT ]−1 Aβ/σ̂ 2 where Cov(x)
χ2 test statistic T = (Aβ̂)T [A(Cov(x)) is a
2 2
consistent estimator of Cov(x) and σ̂ is a consistent
√ estimator of σ . By Chen
and Li (1998), the asymptotic distribution of n(β̂ − cβ) is approximately
Np−1 (0, σ 2[Cov(x)]−1 ).
Theorem 12.4: Li and Duan (1989, p. 1012, 1034-1035). Assume
that the 1D regression model Y x|α + β T x is appropriate. Then under
regularity conditions, the χ2 test of the form Ho: Aβ = 0 that rejects Ho if
T > χ2k,1−α are asymptotically valid.
If ellipsoidal trimming is used and the trimming proportion was not cho-
sen using the response y, then Theorem 12.4 still applies. Using EY plots
to pick M uses y. Make a DD plot using the predictors x. The majority of
the plotted points will often follow some line through the origin. If the DD
plot is used to trim M% of the cases that do not follow the line, as in Figure
11.2, then the following corollary can be used.
Corollary 12.5. Suppose that the trimming proportion M is chosen
CHAPTER 12. 1D REGRESSION 370
without using the response y. Let (yi,M , xi,M ) denote the data that was not
trimmed where i = 1, ..., nM . Then the χ2 test of the form Ho: Aβ = 0 are
asymptotically valid when based on (yi,M , xi,M ).
There is a close relationship between this inference result and the result in
the previous section that showed that variable selection procedures, originally
meant for MLR, can be used for 1D data. Li and Duan (1989) suggests that if
a 1D model Y x|α+βT x is appropriate, then β OLS = cβ for some constant
c. Assume that x is a (p − 1) × 1 vector of nontrivial predictors and that
all models also include a constant. Hence the full model uses p predictors.
Under additional conditions, Li and Duan show that c = 0 and β̂ OLS is
asymptotically normal. For many 1D data sets, the change is SS F statistic
FI can be used to test whether the p − k predictor variables not in I can be
deleted. When the Gaussian multiple linear regression model holds, the p–
value (for Ho: the p − k predictors can be deleted) is approximately p–value
= P (Fp−k,n−p > FI ). That is, if model I is selected before collecting data,
then the FI statistic has an Fp−k,n−p distribution when Ho is true. Theorem
12.4 shows that for many 1D data sets, the FI statistic has an asymptotic
Fp−k,n−p distribution when Ho is true. Hence the p–value ≈ P (Fp−k,n−p >
FI ). The key assumptions for this result are that n is large and no strong
linearities are present in the predictors.
Variable selection with the Cp criterion is closely related to the change in
SS F test. The following results are properties of OLS and hold even if the
data does not follow a 1D model. If the candidate model of xI has k terms
(including the constant), then let
where SSE is the residual sum of squares from the full model and SSE(I) is
the residual sum of squares from the candidate submodel. Then
SSE(I)
Cp (I) = + 2k − n = (p − k)(FI − 1) + k (12.19)
MSE
where MSE is the residual mean square for the full model. Let ESP(I) =
T
α̂I + β̂ I x be the ESP for the submodel and let VI = Y − ESP (I) so that
T
VI,i = Yi − α̂I + β̂ I xi . Let ESP and V denote the corresponding quantities
for the full model. It can be shown that corr(VI , V ) → 1 forces corr(OLS
CHAPTER 12. 1D REGRESSION 371
t2i = FIi .
Using the screen Cp (I) ≤ min(2k, p) suggests that the predictor Xi should
not be deleted if √
|ti | > 2 ≈ 1.414.
More generally, it can be shown that Cp (I) ≤ 2k iff
p
FI ≤ .
p−k
CHAPTER 12. 1D REGRESSION 372
12.6 Complements
An excellent introduction to 1D regression and regression graphics is Cook
and Weisberg (1999a, ch. 18, 19, and 20) and Cook and Weisberg (1999b).
More advanced treatments are Cook (1998a) and Li (2000). Important pa-
pers include Brillinger (1977, 1983), Li and Duan (1989) and Stoker (1986).
Xia, Tong, Li and Zhu (2002) provides a method for single index models
(and multi–index models) that does not need the linearity condition. Formal
testing procedures for the single index model are given by Simonoff and Tsai
(2002).
There are many ways to estimate 1D models, including maximum likeli-
hood for parametric models. The literature for estimating cβ when model
(12.1) holds is growing, and Cook and Li (2002) summarize when compet-
ing methods such as ordinary least squares (OLS), sliced inverse regression
(SIR), principal Hessian directions (PHD), and sliced average variance esti-
mation (SAVE) can fail. All four methods frequently perform well if there
are no strong nonlinearities present in the predictors. Cook and Ni (2005)
provides theory for inverse regression methods such as SAVE. Further in-
formation about these and related methods can be found, for example, in
Brillinger (1977, 1983), Bura and Cook (2001), Chen and Li (1998), Cook
(1998ab, 2000, 2003, 2004), Cook and Critchley (2000), Cook and Li (2004),
Cook and Weisberg (1991, 1999ab), Fung, He, Liu and Shi (2002), Li (1991,
1992, 2000), Satoh and Ohtaki (2004) and Yin and Cook (2002, 2003).
In addition to OLS, specialized methods for 1D models with an unknown
inverse link function (eg models (12.2) and (12.3)) have been developed, and
often the focus is on developing asymptotically efficient methods. See the
references in Cavanagh and Sherman (1998), Delecroix, Härdle and Hris-
tache (2003), Härdle, Hall and Ichimura (1993), Horowitz (1998), Hristache,
CHAPTER 12. 1D REGRESSION 373
Juditsky, Polzehl, and Spokoiny (2001), Stoker (1986), Weisberg and Welsh
(1994) and Xia, Tong, Li and Zhu (2002).
Corollary 12.5 holds for much more general 1D regression methods. If the
trimming is done with a DD plot and the dimension reduction method such
as SIR is performed on the untrimmed data (yi,M , xi,M ), then the inference
that is valid for M = 0 tends to be valid for M > 0.
Several papers have suggested that outliers and strong nonlinearities need
to be removed from the predictors. See Brillinger (1991), Cook (1998a, p.
152), Cook and Nachtsheim (1994), Heng-Hui (2001), Li and Duan (1989,
p. 1011, 1041, 1042) and Li (1991, p. 319). Outlier resistant methods for
general methods such as SIR are less common, but see Gather, Hilker and
Becker (2001, 2002). Trimmed views were introduced by Olive (2002, 2004b).
Li, Cook and Nachtsheim (2004) find clusters, fit OLS to each cluster and
then pool the OLS estimators into a final estimator. This method uses all n
cases while trimmed views gives M% of the cases weight zero. The trimmed
views estimator will often work well when outliers and influential cases are
present.
Section 12.4 follows Olive and Hawkins (2005) closely. The literature on
numerical methods for variable selection in the OLS multiple linear regression
model is enormous, and the literature for other given 1D regression models
is also growing. Li, Cook and Nachtsheim (2005) give an alternative method
for variable selection that can work without specifying the model. Also see,
for example, Claeskins and Hjort (2003), Efron, Hastie, Johnstone and Tib-
shirani (2004), Fan and Li (2001, 2002), Hastie (1987), Lawless and Singhai
(1978), Naik and Tsai (2001), Nordberg (1982), Nott and Leonte (2004), and
Tibshirani (1996). For generalized linear models, forward selection and back-
ward elimination based on the AIC criterion are often used. See Chapter 13,
Agresti (2002, p. 211-217) or Cook and Weisberg (1999a, p. 485, 536-538).
Again, if the variable selection techniques in these papers are successful, then
the estimated sufficient predictors from the full and candidate model should
be highly correlated, and the EE, VV and EY plots will be useful.
The variable selection model with x = (xTS , xTE )T and SP = α + βT x =
α + βTS xS is not the only variable selection model. Burnham and Anderson
(2004) note that for many data sets, the variables can be ordered in decreasing
importance from x1 to xp−1 . The “tapering effects” are such that if n >> p,
then all of the predictors should be used, but for moderate n it is better to
CHAPTER 12. 1D REGRESSION 374
also written as
y x|B T x
where B is the (p − 1) × k matrix
B = [β1 , ..., βk ].
E(x|B T x) = a + CB T x
E(xj |B T x) = aj + cTj B T x
12.7 Problems
12.1. Refer to Definition 12.3 for the Cox and Snell (1968) definition for
residuals, but replace η by β.
a) Find êi if yi = µ + ei and T (Y ) is used to estimate µ.
b) Find êi if yi = xTi β + ei .
c) Find êi if yi = β1 exp[β2(xi − x̄)]ei where the ei are iid exponential(1)
random variables and x̄ is the sample mean of the xi s.
√
d) Find êi if yi = xTi β + ei / wi .
CHAPTER 12. 1D REGRESSION 376
ADJUSTED
k CP R SQUARE R SQUARE RESID SS MODEL VARIABLES
-- ----- -------- -------- --------- ---------------
1 379.8 0.0000 0.0000 37363.2 INTERCEPT ONLY
2 36.0 0.3900 0.3913 22744.6 F
2 113.2 0.3025 0.3039 26007.8 G
2 191.3 0.2140 0.2155 29310.8 E
3 21.3 0.4078 0.4101 22039.9 E F
3 25.0 0.4036 0.4059 22196.7 F H
3 30.8 0.3970 0.3994 22442.0 D F
4 17.5 0.4132 0.4167 21794.9 C E F
4 18.1 0.4125 0.4160 21821.0 E F H
4 18.8 0.4117 0.4152 21850.4 A E F
5 10.2 0.4226 0.4272 21402.3 A E F H
5 10.8 0.4219 0.4265 21427.7 C E F H
5 12.0 0.4206 0.4252 21476.6 A D E F
6 5.7 0.4289 0.4346 21125.8 A C E F H
6 9.3 0.4248 0.4305 21279.1 A C D E F
6 10.3 0.4237 0.4294 21319.9 A B E F H
7 6.3 0.4294 0.4362 21065.0 A B C E F H
7 6.3 0.4294 0.4362 21066.3 A C D E F H
7 7.7 0.4278 0.4346 21124.3 A C E F G H
8 7.0 0.4297 0.4376 21011.8 A B C D E F H
8 8.3 0.4283 0.4362 21064.9 A B C E F G H
8 8.3 0.4283 0.4362 21065.8 A C D E F G H
9 9.0 0.4286 0.4376 21011.8 A B C D E F G H
12.4. The output above is for the Boston housing data from software
that does all subsets variable selection. The full model is a 1D transformation
model with response variable y = CRIM and uses a constant and variables
A, B, C, D, E, F, G and H. (Using log(CRIM) as the response would give an
MLR model.) From this output, what is the best submodel? Explain briefly.
CHAPTER 12. 1D REGRESSION 378
R/Splus Problems
Warning: Use the command source(“A:/rpack.txt”) to download
the programs. See Preface or Section 14.2. Typing the name of the
rpack function, eg trviews, will display the code for the function. Use the
args command, eg args(trviews), to display the needed arguments for the
function.
12.6. Use the following R/Splus commands to make 100 N3 (0, I3 ) cases
and 100 trivariate non-EC cases.
d) After all 10 plots have been looked at the output will show 10 estimated
predictors. The last estimate is the OLS (least squares) view and might look
like
Intercept X1 X2 X3
4.417988 22.468779 61.242178 75.284664
If the OLS view is a good estimated sufficient summary plot, then the
plot created from the command (leave out the intercept)
plot(n3x%*%c(22.469,61.242,75.285),n3x%*%1:3)
should cluster tightly about some line. Your linear combination will be dif-
ferent than the one used above. Using your OLS view, include the plot using
the command above (but with your linear combination) in Word. Was this
plot linear? Did some of the other trimmed views seem to be better that the
OLS view, that is did one of the trimmed views seem to have a smooth mean
function with a smaller variance function than the OLS view?
e) Now type the R/Splus command
lncy <- (ln3x%*%1:3)^3 + 0.1*rnorm(100).
Use the command trviews(ln3x,lncy) to find the best view with a smooth
mean function and the smallest variance function. This view should not be
the OLS view. Include your best view in Word.
f) Get the linear combination from your view, say (94.848, 216.719, 328.444)T ,
and obtain a plot with the command
plot(ln3x%*%c(94.848,216.719,328.444),ln3x%*%1:3).
Include the plot in Word. If the plot is linear with high correlation, then
your EY plot in e) should be good.
12.7. (At the beginning of your R/Splus session, use source(“A:/rpack.txt”)
command (and library(lqs) in R.))
a) Perform the commands
> nx <- matrix(rnorm(300),nrow=100,ncol=3)
> lnx <- exp(nx)
> SP <- lnx%*%1:3
> lnsincy <- sin(SP)/SP + 0.01*rnorm(100)
CHAPTER 12. 1D REGRESSION 380
For parts b), c) and d) below, to make the best trimmed view with
trviews, ctrviews or lmsviews, you may need to use the function twice.
The first view trims 90% of the data, the next view trims 80%, etc. The last
view trims 0% and is the OLS view (or lmsreg view). Remember to advance
the view with the rightmost mouse button (and in R, highlight “stop”). Then
click on the plot and next simultaneously hit Ctrl and c. This makes a copy
of the plot. Then in Word, use the menu commands “Copy>paste.”
b) Find the best trimmed view with OLS and cov.mcd with the following
commands and include the view in Word.
> trviews(lnx,lnsincy)
(With trviews, suppose that 40% trimming gave the best view. Then
instead of using the procedure above b), you can use the command
> essp(lnx,lnsincy,M=40)
to make the best trimmed view. Then click on the plot and next simultane-
ously hit Ctrl and c. This makes a copy of the plot. Then in Word, use the
menu commands “Copy>paste”. Click the rightmost mouse button (and in
R, highlight “stop”) to return the command prompt.)
c) Find the best trimmed view with OLS and (x, S) using the following
commands and include the view in Word. See the paragraph above b).
> ctrviews(lnx,lnsincy)
d) Find the best trimmed view with lmsreg and cov.mcd using the fol-
lowing commands and include the view in Word. See the paragraph above
b).
> lmsviews(lnx,lnsincy)
b) Make sufficient summary plots similar to Figures 12.1 and 12.2 with
the following commands and include both plots in Word.
> plot(SP,ncuby)
> plot(-SP,ncuby)
c) Find the best trimmed view with the following commands (first type
library(lqs) if you are using R). Include the view in Word.
> trviews(nx,ncuby)
You may need to use the function twice. The first view trims 90% of the
data, the next view trims 80%, etc. The last view trims 0% and is the OLS
view. Remember to advance the view with the rightmost mouse button (and
in R, highlight “stop”). Suppose that 40% trimming gave the best view.
Then use the command
to make the best trimmed view. Then click on the plot and next simultane-
ously hit Ctrl and c. This makes a copy of the plot. Then in Word, use the
menu commands “Copy>paste”.
d) To make a plot like Figure 12.5, use the following commands. Let
tem = β̂ obtained from the trviews output. In Example 12.3, tem can be
obtained with the following command.
13.1 Introduction
Generalized linear models are an important class of parametric 1D regres-
sion models that include multiple linear regression, logistic regression and
loglinear regression. Assume that there is a response variable Y and a k × 1
vector of nontrivial predictors x. Before defining a generalized linear model,
the definition of a one parameter exponential family is needed. Let q(y) be
a probability density function (pdf) if Y is a continuous random variable
and let q(y) be a probability mass function (pmf) if Y is a discrete random
variable. Assume that the support of the distribution of Y is Y and that the
parameter space of θ is Θ.
Definition 13.1. A family of pdf’s or pmf’s {q(y|θ) : θ ∈ Θ} is a
1-parameter exponential family if
q(y|θ) = k(θ)h(y) exp[w(θ)t(y)] (13.1)
where k(θ) ≥ 0 and h(y) ≥ 0. The functions h, k, t, and w are real valued
functions.
In the definition, it is crucial that k and w do not depend on y and that
h and t do not depend on θ. The parameterization is not unique since, for
example, w could be multiplied by a nonzero constant m if t is divided by
m. Many other parameterizations are possible. If h(y) = g(y)IY (y), then
usually k(θ) and g(y) are positive, so another parameterization is
q(y|θ) = exp[w(θ)t(y) + d(θ) + S(y)]IY (y) (13.2)
383
CHAPTER 13. GENERALIZED LINEAR MODELS 384
where S(y) = log(g(y)), d(θ) = log(k(θ)), and the support Y does not depend
on θ. Here the indicator function IY (y) = 1 if y ∈ Y and IY (y) = 0, otherwise.
µ(xi ) = g −1 (α + βT xi ). (13.4)
and notice that the value of the parameter θ(xi ) = η(α + βT xi ) depends
on the value of xi . Since the model depends on x only through the linear
predictor α+β T x, a GLM is a 1D regression model. Thus the linear predictor
is also a sufficient predictor.
The following three sections illustrate three of the most important gener-
alized linear models. After selecting a GLM, the investigator will often want
to check whether the model is useful and to perform inference. Several things
to consider are listed below.
i) Show that the GLM provides a simple, useful approximation for the
relationship between the response variable Y and the predictors x.
ii) Estimate α and β using maximum likelihood estimators.
iii) Estimate µ(xi ) = di τ (xi ) or estimate τ (xi ) where the di are known
constants.
iv) Check for goodness of fit of the GLM with an estimated sufficient
summary plot.
v) Check for lack of fit of the GLM (eg with a residual plot).
vii) Check whether Y is independent of x; ie, check whether β = 0.
viii) Check whether a reduced model can be used instead of the full model.
xi) Use variable selection to find a good submodel.
x) Predict Yi given xi .
Yi ≡ Yi |xi = α + β T xi + ei
where ei ∼ N(0, σ 2 ).
When the predictor variables are continuous, the above model is called a
multiple linear regression (MLR) model. When the predictors are categorical,
the above model is called an analysis of variance (ANOVA) model, and when
the predictors are both continuous and categorical, the model is called an
MLR or analysis of covariance model. The MLR model is discussed in detail
in Chapter 5, where the normality assumption and the assumption that σ is
known can be relaxed.
5
0
Y
−5
−10
−10 −5 0 5
SP
5
0
Y
−5
−10
−10 −5 0 5
ESP
0
−1
−2
−10 −5 0 5
ESP
5
0
Y
−5
−10
ESP
exp(α + βT xi )
P (success|xi ) = ρ(xi ) = . (13.5)
1 + exp(α + β T xi )
For the remainder of this section, assume that the binary re-
gression model is of interest. To see that the binary logistic regression
model is a GLM, assume that Y is a binomial(1, ρ) random variable. For a
one parameter family, take a(φ) ≡ 1. Then the pmf of Y is
1 y 1−y 1 ρ
p(y) = ρ (1 − ρ) = (1 − ρ) exp[log( ) y].
y y - ./ 0 1−ρ
-./0 k(ρ)≥0 - ./ 0
h(y)≥0 c(ρ)
−1 exp(α + βT x)
g (α + β x) =T
= ρ(x) = µ(x).
1 + exp(α + βT x)
exp(α + βT x)
P (Y = 1|x) = ρ(x) = .
1 + exp(α + βT x)
and
π1
α = log − 0.5(µ1 − µ0 )T Σ−1 (µ1 + µ0 ).
π2
The discriminant function estimators α̂D and β̂D are found by replacing
the population quantities π1 , π2, µ1 , µ0 and Σ by sample quantities. The
logistic regression (maximum likelihood) estimator also tends to perform well
for this type of data. An exception is when the Y = 0 cases and Y = 1 cases
CHAPTER 13. GENERALIZED LINEAR MODELS 392
can be perfectly or nearly perfectly classified by the ESP. Let the logistic
T
regression ESP = α̂ + β̂ x. Consider the ESS plot of the ESP versus Y . If
the Y = 0 values can be separated from the Y = 1 values by a vertical line (eg
ESP = 0), then there is perfect classification. (If only a few cases need to be
deleted in order for the data set to have perfect classification, then the amount
of “overlap” is small and there is nearly “perfect classification.”) In this
case the maximum likelihood estimator for the logistic regression parameters
(α, β) does not exist because the logistic curve can not approximate a step
function perfectly.
Using Definition 13.4 makes simulation of logistic regression data straight-
forward. Set π0 = π1 = 0.5, Σ = I, and µ0 = 0. Then α ≈ −0.5µT1 µ1
and β = µ1 . The artificial data set used in the following discussion used
β = (1, 1, 1, 0, 0)T and hence α = −1.5. Let Ni be the number of cases where
Y = i for i = 0, 1. For the artificial data, N0 = N1 = 100, and hence the
total sample size n = N1 + N0 = 200.
Again a sufficient summary plot of the sufficient predictor SP = α+βT xi
versus the response variable Yi with the mean function added as a visual aid
can be useful for describing the logistic regression (LR) model. The artificial
data described above was used because the plot can not be used for real data
since α and β are unknown.
Unlike the SSP for multiple linear regression where the mean function
is always the identity line, the mean function in the SSP for LR can take
a variety of shapes depending on the range of the SP. For the LR SSP, the
mean function is
exp(SP )
ρ(SP ) = .
1 + exp(SP )
If the SP = 0 then Y |SP ∼ binomial(1,0.5). If the SP = −5, then Y |SP ∼
binomial(1,ρ ≈ 0.007) while if the SP = 5, then Y |SP ∼ binomial(1,ρ ≈
0.993). Hence if the range of the SP is in the interval (−∞, −5) then the
mean function is flat and ρ(SP ) ≈ 0. If the range of the SP is in the interval
(5, ∞) then the mean function is again flat but ρ(SP ) ≈ 1. If −5 < SP < 0
then the mean function looks like a slide. If −1 < SP < 1 then the mean
function looks linear. If 0 < SP < 5 then the mean function first increases
rapidly and then less and less rapidly. Finally, if −5 < SP < 5 then the
mean function has the characteristic “ESS” shape shown in Figure 13.5.
The estimated sufficient summary plot (ESSP or ESS plot) is a plot of
CHAPTER 13. GENERALIZED LINEAR MODELS 393
1.0
0.8
0.6
Y
0.4
0.2
0.0
−8 −6 −4 −2 0 2 4 6
SP
0.4
0.2
0.0
−6 −4 −2 0 2 4 6
ESP
T
ESP = α̂ + β̂ xi versus Yi with the estimated mean function
exp(ESP )
ρ̂(ESP ) =
1 + exp(ESP )
added as a visual aid. The interpretation of the ESS plot is almost the same
as that of the SSP, but now the SP is estimated by the estimated sufficient
predictor (ESP).
This plot is very useful as a goodness of fit diagnostic. Divide the ESP into
J “slices” each containing approximately n/J cases. Compute the sample
mean = sample proportion of the Y ’s in each slice and add the resulting
step function to the ESS plot. This is done in Figure 13.6 with J = 10
slices. This step function is a simple nonparametric estimator of the mean
function ρ(SP ). If the step function follows the estimated LR mean function
(the logistic curve) closely, then the LR model fits the data well. The plot
of these two curves is a graphical approximation of the goodness of fit tests
described in Hosmer and Lemeshow (2000, p. 147–156).
The deviance test described in Section 13.5 is used to test whether β = 0,
and is the analog of the ANOVA F test for multiple linear regression. If
the LR model is a good approximation to the data but β = 0, then the
predictors x are not needed in the model and ρ̂(xi ) ≡ ρ̂ = Y (the usual
univariate estimator of the success proportion) should be used instead of the
LR estimator
T
exp(α̂ + β̂ xi )
ρ̂(xi ) = T
.
1 + exp(α̂ + β̂ xi )
If the logistic curve clearly fits the step function better than the line Y = Y ,
then Ho will be rejected, but if the line Y = Y fits the step function about
as well as the logistic curve (which should only happen if the logistic curve
is linear with a small slope), then Y may be independent of the predictors.
Figure 13.7 shows the ESS plot when only X4 and X5 are used as predic-
tors for the artificial data, and Y is independent of these two predictors by
construction. It is possible to find data sets that look like Figure 13.7 where
the p–value for the deviance test is very small. Then the LR relationship
is statistically significant, but the investigator needs to decide whether the
relationship is practically significant.
For binary data the Yi only take two values, 0 and 1, and the residuals do
not behave very well. Instead of using residual plots, we suggest using the
binary response plot given in the following definition.
CHAPTER 13. GENERALIZED LINEAR MODELS 395
1.0
0.8
0.6
Y
0.4
0.2
0.0
ESP
Definition 13.5. (Cook 1996): Suppose that the binary response vari-
able Y is conditionally independent of x given the sufficient predictor SP =
α + β T x. Let V be a linear combination of the predictors that is (approx-
imately) uncorrelated with the estimated sufficient predictor ESP. Then a
binary response plot is a plot of the ESP versus V where different plot-
ting symbols are used for Y = 0 and Y = 1.
To make a binary response plot for logistic regression, sliced inverse re-
gression (SIR) can be used to find V . SIR is a regression graphics method
T
and the first SIR predictor β̂SIR1 x is used as the ESP while the second SIR
T
predictor β̂ SIR2x is used as V . (Other regression graphics methods, eg SAVE
or PHD, may provide a better plot, but the first SIR predictor is often highly
T
correlated with the LR ESP α̂ + β̂ x.) After fitting SIR and LR, check that
|corr(SIRESP, LRESP)| ≥ 0.95.
If the LR model holds, then Y is independent of x given the SP, written
Y x|SP.
If the absolute correlation is high, then this conditional independence is ap-
proximately true if the SP is replaced by either the SIR or LR ESP.
CHAPTER 13. GENERALIZED LINEAR MODELS 396
2
1
V
0
−1
−2
−4 −2 0 2
SIRESP
Figure 13.8: This Binary Response Plot Suggests That The Model Is OK
3
2
1
V
0
−1
−2
−3 −2 −1 0 1 2 3
SIRESP
Figure 13.8 shows the binary response plot for the artificial data. The
correlation between the SIR and LR ESP’s was near −1. Hence the slice
symbol density of +’s decreases from nearly 100% in the left of the plot to
0% in the right of the plot. The symbol density is mixed in most of the slices,
suggesting that the LR model is good. For contrast, Figure 13.9 shows the
binary response plot when only X2 and X5 are in the model. Consider the
slice where the ESP is between −2.4 and −1.7. At the bottom and top of
the slice the proportion of +’s is near 1 but in the middle of the slice there
are several 0’s. In the slice where the ESP is between −1.7 and −0.8, the
proportion of +’s increase as one moves from the bottom of the slice to the
top of the slice. Hence there is a large slice from about −2.4 to −0.8 where
the plot does not look good. Although this model is poor, the binary response
plot is inconclusive since only about 20% of the slices are bad. If the bad
slice went from −2.4 to 0.5, the LR model would be bad because more than
25% of the slices would be bad.
Yi ∼ Poisson(µ(xi )).
c(µ) = log(µ).
g −1 (α + βT x) = exp(α + β T x) = µ(x).
Yi ∼ Poisson(exp(α + β T xi ))
µ̂(ESP ) = exp(ESP )
CHAPTER 13. GENERALIZED LINEAR MODELS 400
15
10
Y
5
0
−2 −1 0 1 2
SP
5
0
−2 −1 0 1 2
ESP
added as a visual aid. The interpretation of the EY plot is almost the same
as that of the SSP, but now the SP is estimated by the estimated sufficient
predictor (ESP).
This plot is very useful as a goodness of fit diagnostic. The lowess
curve is a nonparametric estimator of the mean function called a “scatterplot
smoother.” The lowess curve is represented as a jagged curve to distinguish
it from the estimated LLR mean function (the exponential curve) in Figure
13.11. If the lowess curve follows the exponential curve closely (except possi-
bly for the largest values of the ESP), then the LLR model may fit the data
well. A useful lack of fit plot is a plot of the ESP versus the deviance
residuals that are often available from the software.
Warning: For the majority of count data sets where the LLR mean
function is correct, the LLR model is not appropriate but the LLR MLE
is still a consistent estimator of β. The problem is that if Y ∼ P (µ), then
E(Y ) = VAR(Y ) = µ, but for the majority of data sets where E(Y |x) =
µ(x) = exp(SP ), it turns out that VAR(Y |x) > exp(SP ). This phenomenon
is called overdispersion. Adding parametric and nonparametric estimators
of the standard deviation function to the EY plot can be useful. See Cook and
Weisberg (1999a, p. 401-403). Alternatively, if the EY plot looks good and
G2 /(n − 2
√ k − 1) ≈ 1, then the LLR model is likely useful. If G /(n − k − 1) >
1 + 3/ n − k + 1, then a more complicated count model may be needed.
Here the deviance G2 is described in Section 13.5.
The deviance test described in Section 13.5 is used to test whether β = 0,
and is the analog of the ANOVA F test for multiple linear regression. If
the LLR model is a good approximation to the data but β = 0, then the
predictors x are not needed in the model and µ̂(xi ) ≡ µ̂ = Y (the sample
mean) should be used instead of the LLR estimator
T
µ̂(xi ) = exp(α̂ + β̂ xi ).
If the exponential curve clearly fits the lowess curve better than the line
Y = Y , then Ho should be rejected, but if the line Y = Y fits the lowess
curve about as well as the exponential curve (which should only happen if the
exponential curve is approximately linear with a small slope), then Y may be
independent of the predictors. Figure 13.12 shows the ESSP when only X4
and X5 are used as predictors for the artificial data, and Y is independent of
CHAPTER 13. GENERALIZED LINEAR MODELS 402
15
10
Y
5
0
ESP
these two predictors by construction. It is possible to find data sets that look
like Figure 13.12 where the p–value for the deviance test is very small. Then
the LLR relationship is statistically significant, but the investigator needs to
decide whether the relationship is practically significant.
Simple diagnostic plots for the loglinear regression model can be made
using weighted least squares (WLS). To see this, assume that all n of the
counts Yi are large. Then
or
log(Yi ) = α + βT xi + ei
where
Yi
ei = log .
µ(xi )
The error ei does not have zero mean or constant variance, but if µ(xi ) is
large
Yi − µ(xi )
≈ N(0, 1)
µ(xi )
CHAPTER 13. GENERALIZED LINEAR MODELS 403
by the central limit theorem. Recall that log(1 + x) ≈ x for |x| < 0.1. Then,
heuristically,
µ(xi ) + Yi − µ(xi ) Yi − µ(xi )
ei = log ≈ ≈
µ(xi ) µ(xi )
1 Yi − µ(xi ) 1
≈ N 0, .
µ(xi ) µ(xi ) µ(xi )
This suggests that for large µ(xi ), the errors ei are approximately 0 mean
with variance 1/µ(xi ). If the µ(xi ) were known, and all of the Yi were large,
then a weighted least squares of log(Yi ) on xi with weights wi = µ(xi ) should
produce good estimates of (α, β). Since the µ(xi ) are unknown, the estimated
weights wi = Yi could be used. Since P (Yi = 0) > 0, the estimators given in
the following definition are used. Let Zi = Yi if Yi > 0, and let Zi = 0.5 if
Yi = 0.
Definition 13.7. The minimum chi–square estimator of the param-
eters (α, β) in a loglinear regression model are (α̂M , β̂M ), and are found from
the weighted least squares regression of log(Zi ) on xi with weights wi = Zi .
Equivalently,
√ use the ordinary
√ least squares (OLS) regression (without inter-
cept) of Zi log(Zi ) on Zi (1, xTi )T .
The minimum chi–square estimator tends to be consistent if n is fixed
and all n counts Yi increase to ∞ while the loglinear regression maximum
likelihood estimator tends to be consistent if the sample size n → ∞. See
Agresti (2002, p. 611-612). However, the two estimators are often close for
many data sets. This result and the equivalence of the minimum chi–square
estimator to an OLS estimator suggest the following diagnostic plots. Let
(α̃, β̃) be an estimator of (α, β).
Definition 13.8. For a loglinear regression model, a weighted forward
√ √ T √
response plot is a plot of Zi ESP = Zi (α̃ + β̃ xi ) versus Zi log(Zi ).
√ T
The weighted residual plot is a plot of Zi (α̃ + β̃ xi ) versus the WMLR
√ √ T
residuals rW i = Zi log(Zi ) − Zi (α̃ + β̃ xi ).
If the loglinear regression model is appropriate and if the minimum chi–
square estimators are reasonable, then the plotted points in the weighted
forward response plot should follow the identity line. Cases with large WMLR
residuals may not be fit very well by the model. When the counts Yi are
CHAPTER 13. GENERALIZED LINEAR MODELS 404
1.0
8
sqrt(Z) * log(Z)
0.0
WRES
6
4
−1.0
2
−2.0
0
−2 0 2 4 6 8 10 −2 0 2 4 6 8 10
WFIT WFIT
2
8
sqrt(Z) * log(Z)
MWRES
6
1
4
0
2
−1
0
−2 0 2 4 6 8 −2 0 2 4 6 8
MWFIT MWFIT
13.5 Inference
This section gives a very brief discussion of inference for the logistic regression
(LR) and loglinear regression (LLR) models. Inference for these two models
is very similar to inference for the multiple linear regression (MLR) model.
For all three of these models, Y is independent of the k×1 vector of predictors
x = (x1 , ..., xk)T given the sufficient predictor α + β T x:
Y x|α + β T x.
Response = Y
Coefficient Estimates
Number of cases: n
Degrees of freedom: n - k - 1
Pearson X2:
Deviance: D = G^2
-------------------------------------
Binomial Regression
Kernel mean function = Logistic
Response = Status
Terms = (Bottom Left)
Trials = Ones
Coefficient Estimates
Label Estimate Std. Error Est/SE p-value
Constant -389.806 104.224 -3.740 0.0002
Bottom 2.26423 0.333233 6.795 0.0000
Left 2.83356 0.795601 3.562 0.0004
Scale factor: 1.
Number of cases: 200
Degrees of freedom: 197
Pearson X2: 179.809
Deviance: 99.169
that Ho should be rejected, a p–value between 0.01 and 0.07 provides moder-
ate evidence and a p–value less than 0.01 provides strong statistical evidence
that Ho should be rejected. Statistical evidence is not necessarily practical
evidence, and reporting the p–value along with a statement of the strength
of the evidence is more informative than stating that the p–value is less
than some chosen value such as δ = 0.05. Nevertheless, as a homework
convention, use δ = 0.05 if δ is not given.
Investigators also sometimes test whether a predictor Xj is needed in the
model given that the other k − 1 nontrivial predictors are in the model with
a 4 step Wald test of hypotheses:
i) State the hypotheses Ho: βj = 0 Ha: βj = 0.
ii) Find the test statistic zo,j = β̂j /se(β̂j ) or obtain it from output.
iii) The p–value = 2P (Z < −|zoj |) = 2P (Z > |zoj |). Find the p–value from
output or use the standard normal table.
iv) State whether you reject Ho or fail to reject Ho and give a nontechnical
sentence restating your conclusion in terms of the story problem.
If Ho is rejected, then conclude that Xj is needed in the GLM model for
Y given that the other k − 1 predictors are in the model. If you fail to reject
Ho, then conclude that Xj is not needed in the GLM model for Y given that
the other k − 1 predictors are in the model. Note that Xj could be a very
useful GLM predictor, but may not be needed if other predictors are added
to the model.
The Wald confidence interval (CI) for βj can also be obtained from the
output: the large sample 100 (1 − δ) % CI for βj is βˆj ± z1−δ/2 se(β̂j ).
The Wald test and CI tend to give good results if the sample size n
is large. Here 1 − δ refers to the coverage of the CI. Recall that a 90%
CI uses z1−δ/2 = 1.645, a 95% CI uses z1−δ/2 = 1.96, and a 99% CI uses
z1−δ/2 = 2.576.
For a GLM, often 3 models are of interest: the full model that uses all k
of the predictors xT = (xTR, xTO ), the reduced model that uses the r predic-
tors xR , and the saturated model that uses n parameters θ1, ..., θn where
n is the sample size. For the full model the k + 1 parameters α, β1, ..., βk are
estimated while the reduced model has r + 1 parameters. Let lSAT (θ1, ..., θn)
be the likelihood function for the saturated model and let lF U LL(α, β) be the
CHAPTER 13. GENERALIZED LINEAR MODELS 408
be the log likelihood function for the saturated model evaluated at the max-
imum likelihood estimator (MLE) (θ̂1, ..., θ̂n) and let
be the log likelihood function for the full model evaluated at the MLE (α̂, β̂).
Then the deviance
D = G2 = −2(LF U LL − LSAT ).
the estimated sufficient summary plot has been made and that the logistic or
loglinear regression model fits the data well in that the nonparametric step or
lowess estimated mean function follows the estimated model mean function
closely. The deviance test is used to test whether β = 0. If this is the case,
then the predictors are not needed in the GLM model. If Ho : β = 0 is not
rejected, then for loglinear regression the estimator µ̂ = Y should be used
while for logistic regression
n
n
ρ̂ = Yi / mi
i=1 i=1
Total Change
Predictor df Deviance df Deviance
Ones n − 1 = dfo G2o
X1 n−2 1
X2 n−3 1
.. .. .. ..
. . . .
Xk n − k − 1 = dfF U LL G2F U LL 1
-----------------------------------------
Data set = cbrain, Name of Fit = B1
Response = sex
Terms = (cephalic size log[size])
Sequential Analysis of Deviance
Total Change
Predictor df Deviance | df Deviance
Ones 266 363.820 |
cephalic 265 363.605 | 1 0.214643
size 264 315.793 | 1 47.8121
log[size] 263 305.045 | 1 10.7484
CHAPTER 13. GENERALIZED LINEAR MODELS 410
SP = α + β1 x1 + · · · + βk xk = α + β T x = α + β TR xR + β TO xO
where the reduced model uses r of the predictors used by the full model and
xO denotes the vector of k − r predictors that are in the full model but not
the reduced model. For logistic regression, the reduced model is Yi |xRi ∼
independent Binomial(mi, ρ(xRi )) while for loglinear regression the reduced
model is Yi |xRi ∼ independent Poisson(µ(xRi )) for i = 1, ..., n.
CHAPTER 13. GENERALIZED LINEAR MODELS 411
Assume that the ESS plot looks good. Then we want to test Ho : the
reduced model is good (can be used instead of the full model) versus HA :
use the full model (the full model is significantly better than the reduced
model). Fit the full model and the reduced model to get the deviances
G2F U LL and G2RED .
The 4 step change in deviance test is
i) Ho : the reduced model is good HA : use the full model
ii) test statistic G2 (R|F ) = G2RED − G2F U LL
iii) The p–value = P (χ2 > G2 (R|F )) where χ2 ∼ χ2k−r has a chi–square
distribution with k degrees of freedom. Note that k is the number of non-
trivial predictors in the full model while r is the number of nontrivial pre-
dictors in the reduced model. Also notice that k − r = (k + 1) − (r + 1) =
dfRED − dfF U LL = n − r − 1 − (n − k − 1).
iv) Reject Ho if the p–value < δ and conclude that the full model should
be used. If p–value ≥ δ, then fail to reject Ho and conclude that the reduced
model is good.
Interpretation of coefficients: if x1, ..., xi−1, xi+1, ..., xk can be held fixed,
then increasing xi by 1 unit increases the sufficient predictor SP by βi units.
As a special case, consider logistic regression. Let ρ(x) = P (success|x) =
1 − P(failure|x) where a “success” is what is counted and a “failure” is what
is not counted (so if the Yi are binary, ρ(x) = P (Yi = 1|x)). Then the
estimated odds of success is
ρ̂(x) T
Ω̂(x) = = exp(α̂ + β̂ x).
1 − ρ̂(x)
The full model will often contain factors and interaction. If w is a nominal
variable with J levels, make w into a factor by using use J − 1 (indicator or)
dummy variables x1,w , ..., xJ−1,w in the full model. For example, let xi,w = 1 if
w is at its ith level, and let xi,w = 0, otherwise. An interaction is a product
of two or more predictor variables. Interactions are difficult to interpret.
Often interactions are included in the full model, and then the reduced model
without any interactions is tested. The investigator is often hoping that the
interactions are not needed.
A scatterplot of x versus Y is used to visualize the conditional distri-
bution of Y |x. A scatterplot matrix is an array of scatterplots and is used
to examine the marginal relationships of the predictors and response. Place
Y on the top or bottom of the scatterplot matrix. Variables with outliers,
missing values or strong nonlinearities may be so bad that they should not be
included in the full model. Suppose that all values of the variable x are posi-
tive. The log rule says add log(x) to the full model if max(xi )/ min(xi ) > 10.
For the binary logistic regression model, mark the plotted points by a 0 if
Y = 0 and by a + if Y = 1.
To make a full model, use the above discussion and then make an EY
plot to check that the full model is good. The number of predictors in the
CHAPTER 13. GENERALIZED LINEAR MODELS 414
full model should be much smaller than the number of data cases n. Suppose
that the Yi are binary for i = 1, ..., n. Let N1 = Yi = the number of 1’s and
N0 = n−N1 = the number of 0’s. A rough rule of thumb is that the full model
should use no more than min(N0, N1 )/5 predictors and the final submodel
should have r predictor variables where r is small with r ≤ min(N0 , N1)/10.
For loglinear regression, a rough rule of thumb is that the full model should
use no more than n/5 predictors and the final submodel should use no more
than n/10 predictors.
Variable selection, also called subset or model selection, is the search for
a subset of predictor variables that can be deleted without important loss of
information. A model for variable selection for a GLM can be described by
where xI/S denotes the predictors in I that are not in S. Since this is true
regardless of the values of the predictors, βO = 0 if the set of predictors S is
a subset of I. Let (α̂, β̂) and (α̂I , β̂ I ) be the estimates of (α, β) obtained from
fitting the full model and the submodel, respectively. Denote the ESP from
CHAPTER 13. GENERALIZED LINEAR MODELS 415
T
the full model by ESP = α̂ + β̂ xi and denote the ESP from the submodel
by ESP (I) = α̂I + β̂I xIi .
Definition 13.10. An EE plot is a plot of ESP (I) versus ESP .
Variable selection is closely related to the change in deviance test for
a reduced model. You are seeking a subset I of the variables to keep in
the model. The AIC(I) statistic is used as an aid in backward elimination
and forward selection. The full model and the model Imin found with the
smallest AIC are always of interest. Also look for the model Il where the
AIC is the local minimum with the smallest number of nontrivial predictors
(say rl , so that deleting predictors from Il for forward selection or backward
elimination causes AIC to increase). Burnham and Anderson (2004) suggest
that if ∆(I) = AIC(I) − AIC(Imin), then models with ∆(I) ≤ 2 are good,
models with 4 ≤ ∆(I) ≤ 7 are borderline, and models with ∆(I) > 10 should
not be used as the final submodel. Create a full model. The full model has
a deviance at least as small as that of any submodel. The final submodel
should have an EE plot that clusters tightly about the identity line. As a
rough rule of thumb, a good submodel I has corr(ESP (I), ESP ) ≥ 0.95.
Backward elimination starts with the full model with k nontrivial vari-
ables, and the predictor that optimizes some criterion is deleted. Then there
are k − 1 variables left, and the predictor that optimizes some criterion is
deleted. This process continues for models with k − 2, k − 3, ..., 3 and 2
predictors.
Forward selection starts with the model with 0 variables, and the pre-
dictor that optimizes some criterion is added. Then there is 1 variable in
the model, and the predictor that optimizes some criterion is added. This
process continues for models with 2, 3, ..., k − 2 and k − 1 predictors. Both
forward selection and backward elimination result in an (often different) se-
quence of k models {x∗1}, {x∗1, x∗2}, ..., {x∗1, x∗2, ..., x∗k−1}, {x∗1, x∗2, ..., x∗k} = full
model.
All subsets variable selection can be performed with the following
procedure. Compute the ESP of the GLM and compute the OLS ESP found
by the OLS regression of Y on x. Check that |corr(ESP, OLS ESP)| ≥ 0.95.
This high correlation will exist for many data sets. Then perform multiple
linear regression and the corresponding all subsets OLS variable selection
with the Cp(I) criterion. If the sample size n is large and Cp (I) ≤ 2(r + 1)
CHAPTER 13. GENERALIZED LINEAR MODELS 416
the deviance the most. A decrease in deviance less than 4 (if the predictor has
1 degree of freedom) may be troubling in that a bad predictor may have been
added. In practice, the forward selection program may add the variable such
that the submodel I with j nontrivial predictors has a) the smallest AIC(I),
b) the smallest deviance G2 (I) or c) the smallest p–value (preferably from a
change in deviance test but possibly from a Wald test) in the test Ho βi = 0
versus HA βi = 0 where the current model with j terms plus the predictor
xi is treated as the full model (for all variables xi not yet in the model).
Suppose that the full model is good and is stored in M1. Let M2, M3,
M4 and M5 be candidate submodels found after forward selection, backward
elimination, etc. Make a scatterplot matrix of the ESPs for M2, M3, M4,
M5 and M1. Good candidates should have estimated sufficient predictors
that are highly correlated with the full model estimated sufficient predictor
(the correlation should be at least 0.9 and preferably greater than 0.95). For
binary logistic regression, mark the symbols (0 and +) using the response
variable Y .
The final submodel should have few predictors, few variables with large
Wald p–values (0.01 to 0.05 is borderline), a good EY plot and an EE plot
that clusters tightly about the identity line. If a factor has I − 1 dummy
variables, either keep all I − 1 dummy variables or delete all I − 1 dummy
variables, do not delete some of the dummy variables.
13.7 Complements
GLMs were introduced by Nelder and Wedderburn (1972). Books on gen-
eralized linear models (in roughly decreasing order of difficulty) include Mc-
Cullagh and Nelder (1989), Fahrmeir and Tutz (2001), Myers, Montgomery
and Vining (2002) and Dobson (2001). Cook and Weisberg (1999, ch.’s 21-
23) also has an excellent discussion. Texts on categorical data analysis that
have useful discussions of GLM’s include Agresti (2002), Le (1998), Lindsey
(2004), Simonoff (2003) and Powers and Xie (2000) who give econometric
applications. Collett (1999) and Hosmer and Lemeshow (2000) are excellent
texts on logistic regression. See Christensen (1997) for a Bayesian approach
and see Cramer (2003) for econometric applications. Cameron and Trivedi
(1998) and Winkelmann (2000) cover Poisson regression.
The EY plots are also called model checking plots. See Cook and Weisberg
CHAPTER 13. GENERALIZED LINEAR MODELS 418
(1997, 1999a, p. 397, 514, and 541). Cook (1996) introduced the binary
response plot. Also see Cook (1998a, Ch. 5) and Cook and Weisberg (1999a,
section 22.2). Olive and Hawkins (2005) discuss variable selection.
Barndorff-Nielsen (1982) is a very readable discussion of exponential fami-
lies. Also see the webpage (http://www.math.siu.edu/olive/infer.htm). Many
of the distributions in Chapter 3 belong to a 1-parameter exponential family.
A possible method for resistant binary regression is to use trimmed views
but make the ESS plot. This method would work best if x came from an
elliptically contoured distribution. Another possibility is to substitute robust
estimators for the classical estimators in the discrimination estimator.
13.8 Problems
PROBLEMS WITH AN ASTERISK * ARE USEFUL.
13.2∗. Now the data is as in Problem 13.1, but try to estimate the pro-
portion of males by measuring the circumference and the length of the head.
Use the above logistic regression output to answer the following problems.
a) Predict ρ̂(x) if circumference = x1 = 550.0 and length = x2 = 200.0.
b) Perform the 4 step Wald test for Ho : β1 = 0.
c) Perform the 4 step Wald test for Ho : β2 = 0.
13.3∗. A museum has 60 skulls of apes and humans. Lengths of the lower
jaw, upper jaw and face are the explanatory variables. The response variable
is ape (= 1 if ape, 0 if human). Using the output on the previous page,
perform the four step deviance test for whether there is a LR relationship
between the response variable and the predictors.
CHAPTER 13. GENERALIZED LINEAR MODELS 420
Number of cases: 60
Degrees of freedom: 56
Pearson X2: 16.782
Deviance: 13.532
Reduced Model
Response = ape
Coefficient Estimates
Label Estimate Std. Error Est/SE p-value
Constant 8.71977 4.09466 2.130 0.0332
lower jaw -0.376256 0.115757 -3.250 0.0012
upper jaw 0.295507 0.0950855 3.108 0.0019
Number of cases: 60
Degrees of freedom: 57
Pearson X2: 28.049
Deviance: 17.185
13.4∗. Suppose the full model is as in Problem 13.3, but the reduced
model omits the predictor face length. Perform the 4 step change in deviance
test to examine whether the reduced model can be used.
CHAPTER 13. GENERALIZED LINEAR MODELS 421
The following three problems use the possums data from Cook and Weis-
berg (1999a).
13.6∗. Perform the 4 step deviance test for the same model as in Problem
13.5 using the output above.
CHAPTER 13. GENERALIZED LINEAR MODELS 422
13.7∗. Let the reduced model be as in Problem 13.5 and use the output
for the full model be shown above. Perform a 4 step change in deviance test.
B1 B2 B3 B4
df 945 956 968 974
# of predictors 54 43 31 25
# with 0.01 ≤ Wald p-value ≤ 0.05 5 3 2 1
# with Wald p-value > 0.05 8 4 1 0
2
G 892.96 902.14 929.81 956.92
AIC 1002.96 990.14 993.81 1008.912
corr(B1:ETA’U,Bi:ETA’U) 1.0 0.99 0.95 0.90
p-value for change in deviance test 1.0 0.605 0.034 0.0002
13.8∗. The above table gives summary statistics for 4 models considered
as final submodels after performing variable selection. (Several of the predic-
tors were factors, and a factor was considered to have a bad Wald p-value >
0.05 if all of the dummy variables corresponding to the factor had p-values >
0.05. Similarly the factor was considered to have a borderline p-value with
0.01 ≤ p-value ≤ 0.05 if none of the dummy variables corresponding to the
factor had a p-value < 0.01 but at least one dummy variable had a p-value
between 0.01 and 0.05.) The response was binary and logistic regression was
used. The ESS plot for the full model B1 was good. Model B2 was the
minimum AIC model found. There were 1000 cases: for the response, 300
were 0’s and 700 were 1’s.
CHAPTER 13. GENERALIZED LINEAR MODELS 423
a) For the change in deviance test, if the p-value ≥ 0.07, there is little
evidence that Ho should be rejected. If 0.01 ≤ p-value < 0.07 then there is
moderate evidence that Ho should be rejected. If p-value < 0.01 then there
is strong evidence that Ho should be rejected. For which models, if any, is
there strong evidence that “Ho: reduced model is good” should be rejected.
b) For which plot is “corr(B1:ETA’U,Bi:ETA’U)” (using notation from
Arc) relevant?
c) Which model should be used as the final submodel? Explain briefly
why each of the other 3 submodels should not be used.
Arc Problems
The following four problems use data sets from Cook and Weisberg (1999a).
move the slider bar to 1. Move the lowess slider bar until the lowess curve
tracks the exponential curve well. Include the EY plot in Word.
e) Deviance test. From the P2 menu, select Examine submodels and click
on OK. Include the output in Word and perform the 4 step deviance test.
f) Perform the 4 step change of deviance test.
g) EE plot. From Graph&Fit select Plot of. Select P2:Eta’U for the H
box and P1:Eta’U for the V box. Move the OLS slider bar to 1. Click on
the Options popup menu and type “y=x”. Include the plot in Word. Is the
plot linear?
13.12∗. In this problem you will find a good submodel for the possums
data.
a) Activate possums.lsp in Arc with the menu commands
“File > Load > Data > Arcg> possums.lsp.” Scroll up the screen to read
the data description.
b) From Graph&Fit select Fit Poisson response. Select y as the response
and select Acacia, bark, habitat, shrubs, stags and stumps as the predictors.
In Problem 13.11, you showed that this was a good full model.
c) Using what you have learned in class find a good submodel and include
the relevant output in Word.
(Hints: Use forward selection and backward elimination and find a model
that discards a lot of predictors but still has a deviance close to that of the full
model. Also look at the model with the smallest AIC. Either of these models
could be your initial candidate model. Fit this candidate model and look
at the Wald test p–values. Try to eliminate predictors with large p–values
but make sure that the deviance does not increase too much. You may have
several models, say P2, P3, P4 and P5 to look at. Make a scatterplot matrix
of the Pi:ETA’U from these models and from the full model P1. Make the
EE and ESS plots for each model. The correlation in the EE plot should
be at least 0.9 and preferably greater than 0.95. As a very rough guide for
Poisson regression, the number of predictors in the full model should be less
than n/5 and the number of predictors in the final submodel should be less
than n/10.) CONTINUED
CHAPTER 13. GENERALIZED LINEAR MODELS 426
d) Make an EY plot for your final submodel, say P2. From Graph&Fit
select Plot of. Select P2:Eta’U for the H box and y for the V box. From
the OLS popup menu select Poisson and move the slider bar to 1. Move
the lowess slider bar until the lowess curve tracks the exponential curve well.
Include the EY plot in Word.
e) Suppose that P1 contains your full model and P2 contains your final
submodel. Make an EE plot for your final submodel: from Graph&Fit select
Plot of. Select P1:Eta’U for the V box and P2:Eta’U, for the H box. After
the plot appears, click on the options popup menu. A window will appear.
Type y = x and click on OK. This action adds the identity line to the plot.
Also move the OLS slider bar to 1. Include the plot in Word.
f) Using c), d), e) and any additional output that you desire (eg AIC(full),
AIC(min) and AIC(final submodel), explain why your final submodel is good.
Warning: The following problems use data from the book’s web-
page. Save the data files on a disk. Get in Arc and use the menu com-
mands “File > Load” and a window with a Look in box will appear. Click
on the black triangle and then on 3 1/2 Floppy(A:). Then click twice on the
data set name.
13.13∗. (ESS Plot): Activate cbrain.lsp in Arc with the menu commands
“File > Load > 3 1/2 Floppy(A:) > cbrain.lsp.” Scroll up the screen to read
the data description. From Graph&Fit select Fit binomial response. Select
brnweight, cephalic, breadth, cause, size, and headht as predictors, sex as the
response and ones as the number of trials. Perform the logistic regression
and from Graph&Fit select Plot of. Place sex on V and B1:Eta’U on H. From
the OLS popup menu, select Logistic and move the slider bar to 1. From the
lowess popup menu select SliceSmooth and move the slider bar until the fit is
good. Include your plot in Word. Are the slice means (observed proportions)
tracking the logistic curve (fitted proportions) very well?
13.14∗. Suppose that you are given a data set, told the response, and
asked to build a logistic regression model with no further help. In this prob-
lem, we use the cbrain data to illustrate the process.
a) Activate cbrain.lsp in Arc with the menu commands
“File > Load > 1/2 Floppy(A:) > cbrain.lsp.” Scroll up the screen to read
CHAPTER 13. GENERALIZED LINEAR MODELS 427
c) From Graph&Fit select Fit binomial response. Select STA as the re-
sponse and ones as the number of trials. The full model will use every
predictor except ID, LOC and RACE (the latter 2 are replaced by their fac-
tors): select AGE, Bic, CAN, CPR, CRE, CRN, FRA, HRA, INF, {F}LOC ,
PCO, PH, PO2 , PRE , {F}RACE, SER, SEX, SYS and TYP as predictors.
Perform the logistic regression and include the relevant output for testing in
Word. CONTINUED
CHAPTER 13. GENERALIZED LINEAR MODELS 429
d) Make the ESS plot for the full model: from Graph&Fit select Plot of.
Place STA on V and B1:Eta’U on H. From the OLS popup menu, select
Logistic and move the slider bar to 1. From the lowess popup menu select
SliceSmooth and move the slider bar until the fit is good. Include your plot
in Word. Is the full model good?
e) Using what you have learned in class find a good submodel and include
the relevant output in Word. (Hints: Use forward selection and backward
elimination and find a model that discards a lot of predictors but still has
a deviance close to that of the full model. Also look at the model with
the smallest AIC. Either of these models could be your initial candidate
model. Fit this candidate model and look at the Wald test p–values. Try
to eliminate predictors with large p–values but make sure that the deviance
does not increase too much. WARNING: do not delete part of a factor.
Either keep all 2 factor dummy variables or delete all I-1=2 factor dummy
variables. You may have several models, say B2, B3, B4 and B5 to look at.
Make the EE and ESS plots for each model. WARNING: if a factor is in
the full model but not the reduced model, then the EE plot may have I = 3
lines. See part h) below.
f) Make an ESS plot for your final submodel.
g) Suppose that B1 contains your full model and B5 contains your final
submodel. Make an EE plot for your final submodel: from Graph&Fit select
Plot of. Select B1:Eta’U for the V box and B5:Eta’U, for the H box. After
the plot appears, click on the options popup menu. A window will appear.
Type y = x and click on OK. This action adds the identity line to the plot.
Include the plot in Word.
If the EE plot is good and there are one or more factors in the full model
that are not in the final submodel, then the bulk of the data will cluster
tightly about the identity line, but some points may be far away from the
identity line (often lying on some other line) due to the deleted factors.
h) Using e), f), g) and any additional output that you desire (eg AIC(full),
AIC(min) and AIC(final submodel), explain why your final submodel is good.
13.16. In this problem you will examine the museum skull data.
a) Activate museum.lsp in Arc with the menu commands
CHAPTER 13. GENERALIZED LINEAR MODELS 430
“File > Load > 3 1/2 Floppy(A:) > museum.lsp.” Scroll up the screen to
read the data description.
b) From Graph&Fit select Fit binomial response. Select ape as the re-
sponse and ones as the number of trials. Select x5 as the predictor. Perform
the logistic regression and include the relevant output for testing in Word.
c) Make the ESS plot and place it in Word (the response variable is ape
not y). Is the LR model good?
Now you will examine logistic regression when there is perfect classifica-
tion of the sample response variables. Assume that the model used in d)–h)
is in menu B2.
d) From Graph&Fit select Fit binomial response. Select ape as the re-
sponse and ones as the number of trials. Select x3 as the predictor. Perform
the logistic regression and include the relevant output for testing in Word.
e) Make the ESS plot and place it in Word (the response variable is ape
not y). Is the LR model good?
f) Perform the Wald test for Ho : β = 0.
g) From B2 select Examine submodels and include the output in Word.
Then use the output to perform a 4 step deviance test on the submodel used
in part d).
h) The tests in f) and g) are both testing Ho : β = 0 but give different
results. Why are the results different and which test is correct?
13.17. In this problem you will find a good submodel for the credit data
from Fahrmeir and Tutz (2001).
a) Activate credit.lsp in Arc with the menu commands
“File > Load > Floppy(A:) > credit.lsp.” Scroll up the screen to read the
data description. This is a big data set and computations may take several
minutes.
b) Use the menu commands “credit>Make factors” and select x1, x3 , x4, x6,
x7, x8 , x9, x10, x11, x12, x14, x15, x16, and x17. Then click on OK.
c) From Graph&Fit select Fit binomial response. Select y as the response
CHAPTER 13. GENERALIZED LINEAR MODELS 431
and ones as the number of trials. Select {F}x1, x2 , {F}x3, {F}x4, x5, {F}x6,
{F}x7, {F}x8, {F}x9, {F}x10, {F}x11, {F}x12, x13, {F}x14, {F}x15, {F}x16,
{F}x17, x18, x19 and x20 as predictors. Perform the logistic regression and
include the relevant output for testing in Word. You should get 1000 cases,
df = 945, and a deviance of 892.957
d) Make the ESS plot for the full model: from Graph&Fit select Plot
of. Place y on V and B1:Eta’U on H. From the OLS popup menu, select
Logistic and move the slider bar to 1. From the lowess popup menu select
SliceSmooth and move the slider bar until the fit is good. Include your plot
in Word. Is the full model good?
e) Using what you have learned in class find a good submodel and include
the relevant output in Word. (Hints: Use forward selection and backward
elimination and find a model that discards a lot of predictors but still has
a deviance close to that of the full model. Also look at the model with
the smallest AIC. Either of these models could be your initial candidate
model. Fit this candidate model and look at the Wald test p–values. Try
to eliminate predictors with large p–values but make sure that the deviance
does not increase too much. WARNING: do not delete part of a factor.
Either keep all 2 factor dummy variables or delete all I-1=2 factor dummy
variables. You may have several models, say B2, B3, B4 and B5 to look at.
Make the EE and ESS plots for each model. WARNING: if a factor is in
the full model but not the reduced model, then the EE plot may have I = 3
lines. See part h) below.
f) Make an ESS plot for your final submodel.
g) Suppose that B1 contains your full model and B5 contains your final
submodel. Make an EE plot for your final submodel: from Graph&Fit select
Plot of. Select B1:Eta’U for the V box and B5:Eta’U, for the H box. Place
y in the Mark by box. After the plot appears, click on the options popup
menu. A window will appear. Type y = x and click on OK. This action adds
the identity line to the plot. Also move the OLS slider bar to 1. Include the
plot in Word.
h) Using e), f), g) and any additional output that you desire (eg AIC(full),
AIC(min) and AIC(final submodel), explain why your final submodel is good.
CHAPTER 13. GENERALIZED LINEAR MODELS 432
13.18∗. a) This problem uses a data set from Myers, Montgomery and
Vining (2002). Activate popcorn.lsp in Arc with the menu commands
“File > Load > Floppy(A:) > popcorn.lsp.” Scroll up the screen to read the
data description. From Graph&Fit select Fit Poisson response. Use oil, temp
and time as the predictors and y as the response. From Graph&Fit select
Plot of. Select P1:Eta’U for the H box and y for the V box. From the OLS
popup menu select Poisson and move the slider bar to 1. Move the lowess
slider bar until the lowess curve tracks the exponential curve. Include the
EY plot in Word.
b) From the P1 menu select Examine submodels, click on OK and include
the output in Word.
c) Test whether β1 = β2 = β3 = 0.
d) From the popcorn menu, select Transform and select y. Put 1/2 in the
p box and click on OK. From the popcorn menu, select Add a variate and type
yt = sqrt(y)*log(y) in the resulting window. Repeat three times adding the
variates oilt = sqrt(y)*oil, tempt = sqrt(y)*temp and timet = sqrt(y)*time.
From Graph&Fit select Fit linear LS and choose y 1/2, oilt, tempt and timet
as the predictors, yt as the response and click on the Fit intercept box to
remove the check. Then click on OK. From Graph&Fit select Plot of. Select
L2:Fit-Values for the H box and yt for the V box. A plot should appear.
Click on the Options menu and type y = x to add the identity line. Include
the weighted forward response plot in Word.
e) From Graph&Fit select Plot of. Select L2:Fit-Values for the H box and
L2:Residuals for the V box. Include the weighted residual response plot in
Word.
f) For the plot in e), highlight the case in the upper right corner of the
plot by using the mouse to move the arrow just above and to the left the
case. Then hold the rightmost mouse button down and move the mouse to
the right and down. From the Case deletions menu select Delete selection
from data set, then from From Graph&Fit select Fit Poisson response. Use
oil, temp and time as the predictors and y as the response. From Graph&Fit
select Plot of. Select P3:Eta’U for the H box and y for the V box. From
the OLS popup menu select Poisson and move the slider bar to 1. Move the
lowess slider bar until the lowess curve tracks the exponential curve. Include
the EY plot in Word. CONTINUED
CHAPTER 13. GENERALIZED LINEAR MODELS 433
d) Get into SAS, and from the top menu, use the “File> Open” com-
mand. A window will open. Use the arrow in the NE corner of the window
to navigate to “3 1/2 Floppy(A:)”. (As you click on the arrow, you should see
My Documents, C: etc, then 3 1/2 Floppy(A:).) Double click on h10d2.sas.
(Alternatively cut and paste the program into the SAS editor window.) To
execute the program, use the top menu commands “Run>Submit”. An out-
put window will appear if successful. Warning: if you do not have the
two files on A drive, then you need to change the infile command in
h10d2.sas to the drive that you are using, eg change infile “a:cbrain.dat”;
to infile “f:cbrain.dat”; if you are using F drive.
e) To copy and paste relevant output into Word, click on the output
window and use the top menu commands “Edit>Select All” and then the
menu commands “Edit>Copy”.
The model should be good if C(p) ≤ 2k where k = “number in model.”
The only SAS output for this problem that should be included
in Word are two header lines (Number in model, R-square, C(p), Variables
in Model) and the first line with Number in Model = 6 and C(p) = 7.0947.
You may want to copy all of the SAS output into Notepad, and then cut and
paste the relevant two lines of output into Word.
f) Activate cbrain.lsp in Arc with the menu commands
“File > Load > Data > mdata > cbrain.lsp.” From Graph&Fit select Fit
binomial response. Select age = X2, breadth = X6, cephalic = X10, circum
= X9, headht = X4, height = X3, length = X5 and size = X7 as predictors,
sex as the response and ones as the number of trials. This is the full logistic
regression model. Include the relevant output in Word. (A better full model
was used in Problem 13.14.)
g) ESS plot. From Graph&Fit select Plot of. Place sex on V and B1:Eta’U
on H. From the OLS popup menu, select Logistic and move the slider bar
to 1. From the lowess popup menu select SliceSmooth and move the slider
bar until the fit is good. Include your plot in Word. Are the slice means
(observed proportions) tracking the logistic curve (fitted proportions) fairly
well? OVER
h) From Graph&Fit select Fit binomial response. Select breadth = X6,
cephalic = X10, circum = X9, headht = X4, height = X3, and size = X7 as
predictors, sex as the response and ones as the number of trials. This is the
CHAPTER 13. GENERALIZED LINEAR MODELS 436
437
CHAPTER 14. STUFF FOR STUDENTS 438
(http://www.stats.gla.ac.uk/cti/links_stats/journals.html)
(http://www.uni-koeln.de/themen/Statistik/journals.html)
(http://www.stat.uiuc.edu/~he/jourlist.html)
Websites for researchers or research groups can be very useful. Below are
websites for Dr. Rousseeuw’s group, Dr. He, Dr. Rocke, Dr. Croux, Dr.
Hubert’s group and for the University of Minnesota.
(http://www.agoras.ua.ac.be/)
(http://www.stat.uiuc.edu/~he/index.html)
(http://handel.cipic.ucdavis.edu/~dmrocke/preprints.html)
CHAPTER 14. STUFF FOR STUDENTS 440
(http://www.econ.kuleuven.ac.be/public/NDBAE06/)
(http://www.wis.kuleuven.ac.be/stat/robust.html)
(http://www.stat.umn.edu)
The latter website has useful links to software. Arc and R can be down-
loaded from these links. Familiarity with a high level programming
language such as FORTRAN or R/Splus is essential. A very useful R link
is (http://www.r-project.org/#doc).
Finally, a Ph.D. student needs an advisor or mentor and most researchers
will find collaboration valuable. Attending conferences and making your
research available over the internet can lead to contacts.
Some references on research, including technical writing and presenta-
tions, include Becker and Keller-McNulty (1996), Ehrenberg (1982), Hamada
and Sitter (2004), Rubin (2004) and Smith (1997).
> source("A:/rpack.txt")
CHAPTER 14. STUFF FOR STUDENTS 441
The first term in the plot command is always the horizontal axis while the
second is on the vertical axis.
To put a graph in Word, hold down the Cntl and c buttons simultane-
ously. Then select “paste” from the Word Edit menu.
To enter data, open a data set in Notepad or Word. You need to know
the number of rows and the number of columns. Assume that each case is
entered in a row. For example, assuming that the file cyp.lsp has been saved
on your disk from the webpage for this book, open cyp.lsp in Word. It has
76 rows and 8 columns. In R or Splus, write the following command.
cyp <- matrix(scan(),nrow=76,ncol=8,byrow=T)
Then copy the data lines from Word and paste them in R/Splus. If a curser
does not appear, hit enter. The command dim(cyp) will show if you have
entered the data correctly.
Enter the following commands
cypy <- cyp[,2]
cypx<- cyp[,-c(1,2)]
lsfit(cypx,cypy)$coef
to produce the output below.
Intercept X1 X2 X3 X4
205.40825985 0.94653718 0.17514405 0.23415181 0.75927197
X5 X6
-0.05318671 -0.30944144
To check that the data is entered correctly, fit LS in Arc with the re-
sponse variable height and the predictors sternal height, finger to ground,
head length, nasal length, bigonal breadth, and cephalic index (entered in
that order). You should get the same coefficients given by R or Splus.
Making functions in R and Splus is easy.
For example, type the following commands.
mysquare <- function(x){
# this function squares x
r <- x^2
r }
The second line in the function shows how to put comments into functions.
CHAPTER 14. STUFF FOR STUDENTS 443
The following commands are useful for a scatterplot created by the com-
mand plot(x,y).
lines(x,y), lines(lowess(x,y,f=.2))
identify(x,y)
abline(out$coef ), abline(0,1)
The usual arithmetic operators are 2 + 4, 3 − 7, 8 ∗ 4, 8/4, and
2^{10}.
> source("A:/robdata.txt")
> lsfit(belx,bely)
will perform the least squares regression for the Belgian telephone data.
Transferring Data to and from Arc and R or Splus.
For example, suppose that the Belgium telephone data (Rousseeuw and Leroy
1987, p. 26) has the predictor year stored in x and the response number of
calls stored in y in R or Splus. Combine the data into a matrix z and then
use the write.table command to display the data set as shown below. The
sep=’ ’
17 66 14.2
18 67 15.9
19 68 18.2
20 69 21.2
21 70 4.3
22 71 2.4
23 72 2.7073
24 73 2.9
To enter a data set into Arc, use the following template new.lsp.
dataset=new
begin description
Artificial data.
Contributed by David Olive.
end description
begin variables
col 0 = x1
col 1 = x2
col 2 = x3
col 3 = y
end variables
begin data
Next open new.lsp in Notepad. (Or use the vi editor in Unix. Sophisti-
cated editors like Word will often work, but they sometimes add things like
page breaks that do not allow the statistics software to use the file.) Then
copy the data lines from R/Splus and paste them below new.lsp. Then mod-
ify the file new.lsp and save it on a disk as the file belg.lsp. (Or save it in
mdata where mdata is a data folder added within the Arc data folder.) The
header of the new file belg.lsp is shown below.
dataset=belgium
begin description
Belgium telephone data from
Rousseeuw and Leroy (1987, p. 26)
end description
begin variables
CHAPTER 14. STUFF FOR STUDENTS 446
col 0 = case
col 1 = x = year
col 2 = y = number of calls in tens of millions
end variables
begin data
1 50 0.44
. . .
. . .
. . .
24 73 2.9
The file above also shows the first and last lines of data. The header file
needs a data set name, description, variable list and a begin data command.
Often the description can be copied and pasted from source of the data, eg
from the STATLIB website. Note that the first variable starts with Col 0.
To transfer a data set from Arc to R or Splus, select the item
“Display data” from the dataset’s menu. Select the variables you want to
save, and then push the button for “Save in R/Splus format.” You will be
prompted give a file name. If you select bodfat, then two files bodfat.txt and
bodfat.Rd will be created. The file bodfat.txt can be read into either R or Splus
using the read.table command. The file bodfat.Rd saves the documentation
about the data set in a standard format for R.
As an example, the following command was used to enter the body fat
data into Splus. (The mdata folder does not come with Arc. The folder
needs to be created and filled with files from the book’s website. Then the
file bodfat.txt can be stored in the mdata folder.)
The last column of the body fat data consists of the case numbers which
start with 0 in Arc. The second line adds one to each case number.
As another example, use the menu commands
“File>Load>Data>Arcg>forbes.lsp” to activate the forbes data set. From
the Forbes menu, select Display Data. A window will appear. Double click
on Temp and Pressure. Click on Save Data in R/Splus Format and save as
forbes.txt in the folder mdata.
Enter Splus and type the following command.
CHAPTER 14. STUFF FOR STUDENTS 447
forbes<-read.table("C:\\ARC\\DATA\\ARCG\\FORBES.TXT",header=T)
C:\unzipped.
The file
C:\unzipped\dr
contains a folder dr which is the R library. Cut this folder and paste it into
the R library folder. (On my computer, I store the folder rw1011 in the file
C:\Temp.
The folder
C:\Temp\rw1011\library
contains the library packages that came with R.) Open R and type the fol-
lowing command.
library(dr)
Next type help(dr) to make sure that the library is available for use.
CHAPTER 14. STUFF FOR STUDENTS 448
14.3 Projects
Straightforward Projects
• Application
2.2 suggests using Un = n − Ln where Ln = n/2
−
n/4 and
• Read Stigler (1977). This paper suggests a method for comparing new
estimators. Use this method with the two stage estimators TS,n and
TA,n described in Section 2.6.
• Read Anscombe (1961) and Anscombe and Tukey (1963). These papers
suggest graphical methods for checking multiple linear regression and
experimental design methods that were the “state of the art” at the
time. What graphical procedures did they use and what are the most
important procedures that were not suggested?
• Read Bentler and Yuan (1998) and Cattell (1966). These papers use
scree plots to determine how many eigenvalues of the covariance ma-
trix are nonzero. This topic is very important for dimension reduction
methods such as principal components.
CHAPTER 14. STUFF FOR STUDENTS 449
• The simulation study in Section 4.6 suggests that TS,n does not work
well on exponential data. Find a coarse grid so that TS,n works well
normal and exponential data. Illustrate with a simulation study.
• Examine via simulation how the graphical method for assessing variable
selection complements numerical methods. Find at least two data sets
where deleting one case changes the model selected by a numerical
variable selection method such as Cp .
• Find some benchmark multiple linear regression outlier data sets. Fit
OLS, L1 and M-estimators from R/Splus. Are any of the M-estimators
as good as L1 ? (Note: l1fit is in Splus but not in R.)
• Compare lmsreg and the MBA estimator on real and simulated mul-
tiple linear regression data.
• Find some benchmark multiple linear regression outlier data sets. Fit
robust estimators such as ltsreg from R/Splus, but do not use lmsreg.
Are any of the robust estimators as good as the MBA estimator?
• There are several papers that give tests or diagnostics for linearity.
Find a data set from such a paper and find the fitted values from some
nonparametric method. Plot these fitted values versus the fitted values
from a multiple linear regression such as OLS. What should this plot
look like? How can the forward response plot and trimmed views be
used as a diagnostic for linearity? See Hawkins and Olive (2002, p.
158).
CHAPTER 14. STUFF FOR STUDENTS 450
• Using ESP to Search for the Missing Link: Compare trimmed views
which uses OLS and cov.mcd with another regression–MLD combo.
There are 8 possible projects: i) OLS–MBA, ii) OLS–Classical (use
ctrviews), iii) SIR–cov.mcd (sirviews), iv) SIR–MBA, v) SIR–class-
ical, vi) lmsreg–cov.mcd (lmsviews), vii) lmsreg–MBA, and viii) lmsreg
–classical. Do Problem 12.7ac (but just copy and paste the best view
instead of using the essp(nx,ncuby,M=40) command) with both your
estimator and trimmed views. Try to see what types of functions
work for both estimators, when trimmed views is better and when the
procedure i)–viii) in better. If you can invent interesting 1D functions,
do so.
Harder Projects
• The Super Duper Outlier Scooper for MLR: Write R/Splus functions
to compute the two estimators given by Theorem 8.7. Compare these
estimators with lmsreg and ltsreg on real and simulated data.
• The Super Duper Outlier Scooper for Multivariate Location and Disper-
sion: Consider the modified MBA estimator for multivariate location
and dispersion given in Problem 10.17. This MBA estimator uses 8
starts using 0%, 50%, 60%, 70%, 80%, 90%, 95% and 98% trimming of
the cases closest to √ the coordinatewise median in Euclidean distance.
The estimator is n consistent on elliptically contoured distributions
with 2nd moments. For small data sets the cmba2 function can fail
because the covariance estimator applied to the closest 2% cases to
the coordinatewise median is singular. Modify the function so that it
works well on small data sets. Then consider the following proposal
that may make the estimator asymptotically equivalent to the classi-
cal estimator when the data are from a multivariate normal (MVN)
distribution. The attractor corresponding to 0% trimming is the DGK
estimator (µ̂0, Σ̂0). Let (µ̂T , Σ̂T ) = (µ̂0 , Σ̂0 ) if det(Σ̂0 ) ≤ det(Σ̂M ) and
(µ̂T , Σ̂T ) = (µ̂M , Σ̂M ) otherwise where (µ̂M , Σ̂M ) is the attractor cor-
responding to M% trimming. Then make the DD plot of the classical
Mahalanobis distances versus the distances corresponding to (µ̂T , Σ̂T )
for M = 50, 60, 70, 80, 90, 95 and 98. If all seven DD plots “look good”
then use the classical estimator. The resulting estimator will be asymp-
totically equivalent to the classical estimator if P(all seven DD plots
“look good”) goes to one as n → ∞. We conjecture that all seven plots
will look good because if n is large and the trimmed attractor “beats”
the DGK estimator, then the plot will look good. Also if the data is
MVN but not spherical, then the DGK estimator will almost always
“beat” the trimmed estimator, so all 7 plots will be identical.
• The TV estimator for MLR has a good combination of resistance and
theory. Consider the following modification to make the method asymp-
totically equivalent to OLS when the Gaussian model holds: if each
trimmed view “looks good,” use OLS. The method is asymptotically
equivalent to OLS if the probability P(all 10 trimmed views look good)
goes to one as n → ∞. Rousseeuw and Leroy (1987, p. 128) shows
that if the predictors are bounded, then the ith residual ri converges
CHAPTER 14. STUFF FOR STUDENTS 453
• For nonlinear regression models of the form yi = m(xi , β)+ei , the fitted
values are ŷi = m(xi , β̂) and the residuals are ri = yi − ŷi . The points
in the FY plot of the fitted values versus the response should follow
the identity line. The TV estimator would make FY and residual plots
for each of the trimming proportions. The MBA estimator with the
median squared residual criterion can also be used for many of these
models.
• Econometrics project: Suppose that the MLR model holds but Var(e) =
σ 2Σ and Σ = U U where U is known and nonsingular. Show that
U −1 Y = U −1 Xβ + U −1 e, and the TV and MBA estimators can be
applied to Ỹ = U −1 Y and X̃ = U −1 X provided that OLS is fit
without an intercept.
• Suppose that the data set contains missing values. Code the missing
value as ±99999+ rnorm(1). Run a robust procedure on the data. The
idea is that the case with the missing value will be given weight zero if
the variable is important, and the variable will be given weight zero if
the case is important. See Hawkins and Olive (1999b).
• Read Stefanski and Boos (2002). One of the most promising uses of
M-estimators is as generalized estimating equations.
• Robust sequential procedures do not seem to work very well. Try using
analogs of the two stage trimmed means. An ad hoc procedure that
has worked very well is to clean the data using the median and mad
at each sample size. Then apply the classical sequential method and
stopping rule to the cleaned data. This procedure is rather expensive
since the median and mad need to be recomputed with each additional
observation until the stopping rule ends data collection. Another idea
is to examine similar methods in the quality control literature.
• Apply the Cook and Olive (2001) graphical procedure for response
transformations described in Section 5.1 with the power family replaced
by the Yeo and Johnson (2000) family of transformations.
• Prove that the LTA estimator is are consistent, and prove that the LTA
and LTS estimators are Op (n−1/2 ). Prove Conjecture 7.1.
These results are in the folklore but have not been shown outside of
the location model. Mašı̈ček (2004) proved that LTS is consistent.
e)
Sn2 (d1 , ..., dn) 7.0156
VSW = = = 19.4877,
([Un − Ln ]/n)2 ( 8−2
10
)2
so
SE(Tn ) = VSW /n = 19.4877/10 = 1.3960.
2.14 a) Ln = n/2
− n/4 = [5/2] − 5/4 = 2 − 2 = 0.
Un = n − Ln = 5 − 0 = 5.
p = Un − Ln − 1 = 5 − 0 − 1 = 4.
SE(MED(n)) = (Y(Un ) − Y(Ln +1) )/2 = (8 − 2)/2 = 3.
b) Ln = n/4
= 1
= 1.
Un = n − Ln = 5 − 1 = 4.
p = Un − Ln − 1 = 4 − 1 − 1 = 2.
Tn = (3 + 5 + 6)/3 = 4.6667.
The d s are 3 3 5 6 6.
( di )/n = 4.6
( d2i − n(d)2 )/(n − 1) = (115 − 5(4.6)2 )/4 = 9.2/4 = 2.3.
The R/Splus functions for Problems 4.10–4.14 are available from the
text’s website file rpack.txt and should have been entered into the computer
using the source(“A:/rpack.txt”) as described on p. 438.
4.13b i) Coverages should be near 0.95. The lengths should be about 4.3
for n = 10, 4.0 for n = 50 and 3.96 for n = 100.
ii) Coverage should be near 0.78 for n = 10 and 0 for n = 50, 100. The
lengths should be about 187 for n = 10, 173 for n = 50 and 171 for n = 100.
(It can be shown that the expected length for large n is 169.786.)
Chapter 5
5.1 a) 7+ βXi
b) b = (Yi − 7)Xi / Xi2
c) The second derivative = 2 Xi2 > 0.
5.4 Fo = 0.904, p–value > 0.1, fail to reject Ho, so the reduced model is
good
5.5 a) 25.970
b) Fo = 0.600, p–value > 0.5, fail to reject Ho, so the reduced model is
good
2
5.6
2 a) b 3 = X3i (Yi − 10 − 2X2i )/ X3i . The second partial derivative
= X3i > 0.
5.9 a) (1.229, 3.345)
b) (1.0825, 3.4919)
5.11 c) Fo = 265.96, pvalue = 0.0, reject Ho, there is a MLR relationship
between the response variable height and the predictors sternal height and
finger to ground.
5.13 No, the relationship should be linear.
5.14 No, since 0 is in the CI. X could be a very useful predictor for Y ,
eg if Y = X 2 .
5.16 The model uses constant, finger to ground and sternal height. (You
can tell what the variable are by looking at which variables are deleted.)
5.17 Use L3. L1 and L2 have more predictors and higher Cp than L3
CHAPTER 14. STUFF FOR STUDENTS 462
5.24 This problem has the student reproduce Example 5.1. Hence log(Y )
is the appropriate response transformation.
5.25 Plots b), c) and e) suggest that log(ht) is needed while plots d), f)
and g) suggest that log(ht) is not needed. Plots c) and d) show that the
residuals from both models are quite small compared to the fitted values.
Plot d) suggests that the two models produce approximately the same fitted
values. Hence if the goal is prediction, the expensive log(ht) measurement
does not seem to be needed.
5.26 h) The submodel is ok, but the forward response and residual plots
found in f) for the submodel do not look as good as those for the full model
found in d). Since the submodel residuals do not look good, more terms are
probably needed in the model.
CHAPTER 14. STUFF FOR STUDENTS 463
5.29 b) Forward selection gives constant, (size)1/3, age, sex, breadth and
cause.
c) Backward elimination gives constant, age, cause, cephalic, headht,
length and sex.
d) Forward selection is better because it has fewer terms and a smaller
Cp .
e) The variables are highly correlated. Hence backward elimination quickly
eliminates the single best predictor (size)1/3 and can not get a good model
that only has a few terms.
f) Although the model in c) could be used, a better model uses constant,
age, sex and (size)1/3.
j) The FF and RR plots are good and so are the forward response and
residual plots if you ignore the good leverage points corresponding to the 5
babies.
Chapter 6
6.1 b) Masking since 3 outliers are good cases with respect to Cook’s
distances.
c) and d) usually the MBA residuals will be large in magnitude, but for
some students MBA, ALMS and ALTS will be highly correlated.
6.4. a) The AR(2) model has the highest correlation with the response
and is the simplest model. The top row of the scatterplot matrix gives the
FY plots for the 5 different estimators.
b) The AR(11) and AR(12) fits are highly correlated as are the SE-
TAR(2,7,2) and SETAR(2,5,2) fits.
6.6 The response Y with a constant and X3 , X7 , X13 and X14 as predictors
is a good submodel. (A competitor would delete X13 but then the residual
plot is not as good.)
6.8 The response Y with a constant, X2 and X5 as predictors is a good
submodel. One outlier is visible in the residual plot. (A competitor would
also use X3 .)
6.9 The submodel using a constant and X1 is ok although the residual
plot does not look very good.
6.13 The model using log(X3 ), log(X4 ), log(X6 ), log(X11 ), log(X13 ) and
CHAPTER 14. STUFF FOR STUDENTS 464
log(X14 ) plus a constant has a good FF plot but more variables may be
needed to get a good RR plot.
6.14 There are many good models including the submodel that uses
Y = log(BigMac) and a constant, log(BusFare) log(EngSal), log(Service),
log(TeachSal) and log(TeachTax) as predictors.
6.16 e) R2 went from 0.978 with outliers to R2 = 0.165 without the
outliers. (The low value of R2 suggests that the MLR relationship is weak,
not that the MLR model is bad.)
Chapter 7
7.4 b) The line should go through the left and right cluster but not
through the middle cluster of outliers.
c) The identity line should NOT PASS through the cluster of outliers
with Y near 0 and the residuals corresponding to these outliers should be
large in magnitude.
7.5 e) Usually the MBA esitmator based on the median squared residual
will pass through the outliers with the MBA LATA estimator gives zero
weight to the outliers (so that the outliers are large in magnitude).
Chapter 8
8.1. Approximately 2 nδ f(0) cases have small errors.
Chapter 10
10.1 a) X2 ∼ N(100, 6).
b)
X1 49 3 −1
∼ N2 , .
X3 17 −1 4
c) X1 X4 and X3 X4 .
d)
Cov(X1, X3 ) −1
ρ(X1 , X2 ) = = √ √ = −0.2887.
VAR(X1 )VAR(X3 ) 3 4
10.2 a) Y |X ∼ N(49, 16) since Y X. (Or use E(Y |X) = µY +
−1
Σ12Σ22 (X − µx ) = 49 + 0(1/25)(X − 100) = 49 and VAR(Y |X) = Σ11 −
Σ12Σ−1
22 Σ21 = 16 − 0(1/25)0 = 16.)
CHAPTER 14. STUFF FOR STUDENTS 465
10.4 The proof is identical to that given in Example 10.2. (In addition,
it is fairly simple to show that M1 = M2 ≡ M. That is, M depends on Σ
but not on c or g.)
10.6 a) Sort each column, then find the median of each column. Then
MED(W ) = (1430, 180, 120)T .
b) The sample mean of (X1 , X2 , X3 )T is found by finding the sample mean
of each column. Hence x = (1232.8571, 168.00, 112.00)T .
10.11 ΣB = E[E(X|B T X)X T B)] = E(M B B T XX T B) = M B B T ΣB.
Hence M B = ΣB(B T ΣB)−1 .
10.15 The 4 plots should look nearly identical with the five cases 61–65
appearing as outliers.
10.16 Not only should none of the outliers be highlighted, but the high-
lighted cases should be ellipsoidal.
10.17 Answers will vary since this is simulated data, but should get gam
near 0.4, 0.3, 0.2 and 0.1 as p increases from 2 to 20.
Chapter 11
11.2 b Ideally the answer to this problem and Problem 11.3b would be
nearly the same, but students seem to want correlations to be very high and
use n to high. Values of n around 50, 60 and 80 for p = 2, 3 and 4 should be
enough.
11.3 b Values of n should be near 50, 60 and 80 for p = 2, 3 and 4.
11.4 This is simulated data, but for most plots the slope is near 2.
11.8 The identity line should NOT PASS through the cluster of out-
liers with Y near 0. The amount of trimming seems to vary some with the
computer (which should not happen unless there is a bug in the tvreg2 func-
tion or if the computers are using different versions of cov.mcd), but most
students liked 70% or 80% trimming.
CHAPTER 14. STUFF FOR STUDENTS 466
Chapter 12
12.1.
a) êi = yi − T (Y ).
b) êi = yi − xTi β̂.
c)
yi
êi = .
β̂1 exp[β̂2(xi − x̄)]
√
d) êi = wi (yi − xTi β̂).
12.2.
a) Since y is a (random) scalar and E(w) = 0, Σx,y = E[(x − E(x))(y −
E(y))T ] = E[w(y − E(y))] = E(wy) − E(w)E(y) = E(wy).
b) Using the definition of z and r, note that y = m(z) + e and
w = r + (Σx β)βT w. Hence E(wy) = E[(r + (Σx β)βT w)(m(z) + e)] =
E[(r + (Σx β)βT w)m(z)] + E[r + (Σx β)β T w]E(e) since e is independent of
x. Since E(e) = 0, the latter term drops out. Since m(z) and β T wm(z) are
(random) scalars, E(wy) = E[m(z)r] + E[β T w m(z)]Σxβ.
c) Using result b), Σ−1 −1 −1
x Σx,y = Σx E[m(z)r] + Σx E[β w m(z)]Σxβ =
T
E[βT w m(z)]Σ−1 −1 −1
x Σx β + Σx E[m(z)r] = E[β w m(z)]β + Σx E[m(z)r] and
T
g) The slice means follow the logistic curve fairly well with 8 slices.
i) The EE plot is linear.
j) The slice means follow the logistic curve fairly well with 8 slices.
BIBLIOGRAPHY 470
12. Appa, G.M., and Land, A.H. (1993), “Comment on ‘A Cautionary Note
on the Method of Least Median of Squares’ by Hettmansperger, T.P.
and Sheather, S.J.,” The American Statistician, 47, 160-162.
18. Atkinson, A., and Riani, R. (2000), Robust Diagnostic Regression Anal-
ysis, Springer-Verlag, NY.
19. Atkinson, A., Riani, R., and Cerioli, A. (2004), Exploring Multivariate
Data with the Forward Search, Springer-Verlag, NY.
20. Atkinson, A.C., and Weisberg, S. (1991), “Simulated Annealing for the
Detection of Multiple Outliers Using Least Squares and Least Median
of Squares Fitting,” in Directions in Robust Statistics and Diagnostics,
Part 1, eds. Stahel, W., and Weisberg, S., Springer-Verlag, NY, 7-20.
21. Bai, Z.D., and He, X. (1999), “Asymptotic Distributions of the Maxi-
mal Depth Estimators for Regression and Multivariate Location,” The
Annals of Statistics, 27, 1616-1637.
23. Barnett, V., and Lewis, T. (1994), Outliers in Statistical Data, 3rd ed.,
John Wiley and Sons, NY.
BIBLIOGRAPHY 472
24. Barrett, B.E., and Gray, J.B. (1992), “Diagnosing Joint Influence in
Regression Analysis,” in the American Statistical 1992 Proceedings of
the Computing Section, 40-45.
25. Barrodale, I., and Roberts, F.D.K. (1974), “Algorithm 478 Solution of
an Overdetermined System of Equations in the l1 Norm [F 4],” Com-
munications of the ACM, 17, 319-320.
29. Becker, R.A., Chambers, J.M., and Wilks, A.R. (1988), The New S
Language A Programming Environment for Data Analysis and Graph-
ics, Wadsworth and Brooks/Cole, Pacific Grove, CA.
32. Belsley, D.A., Kuh, E., and Welsch, R.E. (1980), Regression Diagnos-
tics: Identifying Influential Data and Sources of Collinearity, John Wi-
ley and Sons, NY.
33. Bentler, P.M., and Yuan K.H. (1998), “Tests for Linear Trend in the
Smallest Eigenvalues of the Correlation Matrix,” Psychometrika, 63,
131-144.
35. Bickel, P.J. (1965), “On Some Robust Estimates of Location,” The
Annals of Mathematical Statistics, 36, 847-858.
BIBLIOGRAPHY 473
36. Bickel, P.J. (1975), “One-Step Huber Estimates in the Linear Model,”
Journal of the American Statistical Association, 70, 428-434.
37. Bickel, P.J., and Doksum, K.A. (1977), Mathematical Statistics: Basic
Ideas and Selected Topics, Holden-Day, San Francisco, CA.
38. Bloch, D.A., and Gastwirth, J.L. (1968), “On a Simple Estimate of
the Reciprocal of the Density Function,” The Annals of Mathematical
Statistics, 39, 1083-1085.
40. Bogdan, M. (1999), “Data Driven Smooth Tests for Bivariate Normal-
ity,” Journal of Multivariate Analysis, 68, 26-53.
41. Bowman, K.O., and Shenton, L.R. (1988), Properties of Estimators for
the Gamma Distribution, Marcel Dekker, NY.
44. Box, G.E.P., and Cox, D.R. (1964), “An Analysis of Transformations,”
Journal of the Royal Statistical Society, B, 26, 211-246.
45. Branco, J.A., Croux, C., Filzmoser, P., and Oliviera, M.R. (2005), “Ro-
bust Canonical Correlations: a Comparative Study,” Computational
Statistics, To Appear.
49. Brockwell, P.J., and Davis, R.A. (1991), Time Series: Theory and
Methods, Springer–Verlag, NY.
50. Broffitt, J.D. (1974), “An Example of the Large Sample Behavior of
the Midrange,” The American Statistician, 28, 69-70.
52. Buja, A., Hastie, T., and Tibshirani, R. (1989), “Linear Smoothers and
Additive Models,” The Annals of Statistics, 17, 453-555.
53. Bura, E., and Cook, R.D., (2001), “Estimating the Structural Dimen-
sion of Regressions Via Parametric Inverse Regression,” Journal of the
Royal Statistical Society, B, 63, 393-410.
57. Butler, R.W., Davies, P.L., and Jhun, M. (1993), “Asymptotics for the
Minimum Covariance Determinant Estimator,” The Annals of Statis-
tics, 21, 1385-1400.
59. Cambanis, S., Huang, S., and Simons, G. (1981) “On the Theory of El-
liptically Contoured Distributions,” Journal of Multivariate Analysis,
11, 368-385.
60. Cameron, A.C., and Trivedi, P.K. (1998), Regression Analysis of Count
Data, Cambridge University Press, Cambridge, UK.
61. Carroll, R.J., and Welsh, A.H. (1988), “A Note on Asymmetry and
Robustness in Linear Regression,” The American Statistician, 42, 285-
287.
62. Casella, G., and Berger, R.L. (2002), Statistical Inference, 2nd ed.,
Duxbury, Belmont, CA.
64. Cattell, R.B. (1966), “The Scree Test for the Number of Factors,” Mul-
tivariate Behavioral Research, 1, 245-276.
65. Cavanagh, C., and Sherman, R.P. (1998), “Rank Estimators for Mono-
tonic Index Models,” Journal of Econometrics, 84, 351-381.
67. Chambers, J.M., Cleveland, W.S., Kleiner, B., and Tukey, P. (1983),
Graphical Methods for Data Analysis, Duxbury Press, Boston.
68. Chang, W.H., McKean, J.W., Naranjo, J.D., and Sheather, S.J. (1999),
“High-Breakdown Rank Regression,” Journal of the American Statis-
tical Association, 94, 205-219.
69. Chatterjee, S., and Hadi, A.S. (1988), Sensitivity Analysis in Linear
Regression, John Wiley and Sons, NY.
70. Chen, C.H., and Li, K.C. (1998), “Can SIR be as Popular as Multiple
Linear Regression?,” Statistica Sinica, 8, 289-316.
71. Chen, J., and Rubin, H. (1986), “Bounds for the Difference Between
Median and Mean of Gamma and Poisson Distributions,” Statistics and
Probability Letters, 4, 281-283.
BIBLIOGRAPHY 476
75. Claeskins, G., and Hjort, N.L. (2003), “The Focused Information Cri-
terion,” (with discussion), Journal of the American Statistical Associ-
ation, 98, 900-916.
78. Cohen, A. C., and Whitten, B.J. (1988), Parameter Estimation in Re-
liability and Life Span Models, Marcel Dekker, NY.
83. Cook, R.D. (1996), “Graphics for Regressions with Binary Response,”
Journal of the American Statistical Association, 91, 983-992.
84. Cook, R.D. (1998a), Regression Graphics: Ideas for Studying Regres-
sion Through Graphics, John Wiley and Sons, NY.
BIBLIOGRAPHY 477
86. Cook, R.D. (2000), “SAVE: A Method for Dimension Reduction and
Graphics in Regression,” Communications in Statistics Theory and
Methods, 29, 2109-2121.
89. Cook, R.D., and Critchley, F. (2000), “Identifying Outliers and Regres-
sion Mixtures Graphically,” Journal of the American Statistical Asso-
ciation, 95, 781-794.
93. Cook, R.D., Hawkins, D.M., and Weisberg, S. (1993), “Exact Iterative
Computation of the Robust Multivariate Minimum Volume Ellipsoid
Estimator,” Statistics and Probability Letters, 16, 213-218.
94. Cook, R.D., and Lee, H. (1999), “Dimension Reduction in Binary Re-
sponse Regression,” Journal of the American Statistical Association,
94, 1187-1200.
BIBLIOGRAPHY 478
95. Cook, R.D., and Li, B. (2002), “Dimension Reduction for Conditional
Mean in Regression,” The Annals of Statistics, 30, 455-474.
96. Cook, R.D., and Li, B. (2004), “Determining the Dimension of Iterative
Hessian Transformation,” The Annals of Statistics, 32, 2501-2531.
98. Cook, R.D., and Ni, L. (2005), “Sufficient Dimension Reduction via
Inverse Regression: A Minimum Discrepancy Approach,” Journal of
the American Statistical Association, 100, 410-428.
99. Cook, R.D., and Olive, D.J. (2001), “A Note on Visualizing Response
Transformations in Regression,” Technometrics, 43, 443-449.
100. Cook, R.D., and Wang, P.C. (1983), “Transformations and Influential
Cases in Regression,” Technometrics, 25, 337-343.
101. Cook, R.D., and Weisberg, S. (1982), Residuals and Influence in Re-
gression, Chapman & Hall, London.
102. Cook, R.D., and Weisberg, S. (1991), “Comment on ‘Sliced Inverse Re-
gression for Dimension Reduction’ by K.C. Li,” Journal of the Ameri-
can Statistical Association, 86, 328-332.
104. Cook, R.D., and Weisberg, S. (1997), “Graphs for Assessing the Ade-
quacy of Regression Models,” Journal of the American Statistical As-
sociation, 92, 490-499.
107. Cooke, D., Craven, A.H., and Clarke, G.M. (1982), Basic Statistical
Computing, Edward Arnold Publishers, London.
BIBLIOGRAPHY 479
109. Cox, D.R. (1972), “Regression Models and Life-Tables,” Journal of the
Royal Statistical Society, B, 34, 187-220.
111. Cramér, J.S. (2003), Logit Models from Economics and Other Fields,
Cambridge University Press, Cambridge, UK.
112. Croux, C., Dehon, C., Rousseeuw, P.J., and Van Aelst, S. (2001), “Ro-
bust Estimation of the Conditional Median Function at Elliptical Mod-
els”, Statistics and Probability Letters, 51, 361-368.
113. Croux, C., and Haesbroeck, G. (2003), “Implementing the Bianco and
Yohai Estimator for Logistic Regression,” Computational Statistics and
Data Analysis, 44, 273-295.
117. Daniel, C., and Wood, F.S. (1980), Fitting Equations to Data, 2nd ed.,
John Wiley and Sons, NY.
119. David, H.A. (1981), Order Statistics, 2nd ed., John Wiley and Sons,
NY.
BIBLIOGRAPHY 480
120. David, H.A. (1995), “First (?) Occurrences of Common Terms in Math-
ematical Statistics,” The American Statistician, 49, 121-133.
122. Davies, L., and Gather, U. (1993), “The Identification of Multiple Out-
liers,” Journal of the American Statistical Association, 88, 782-792.
125. Davies, P.L. (1993), “Aspects of Robust Linear Regression,” The An-
nals of Statistics, 21, 1843-1899.
126. deCani, J.S, and Stine, R.A. (1986), “A Note on Deriving the Infor-
mation Matrix for a Logistic Distribution,” The American Statistician,
40, 220-222.
128. Delecroix, M., Härdle, W., and Hristache, M. (2003), “Efficient Estima-
tion in Conditional Single-Index Regression,” Journal of Multivariate
Analysis, 86, 213-226.
130. Devlin, S.J., Gnanadesikan, R., and Kettenring, J.R. (1975), “Ro-
bust Estimation and Outlier Detection with Correlation Coefficients,”
Biometrika, 62, 531-545.
131. Devlin, S.J., Gnanadesikan, R., and Kettenring, J.R. (1981), “Robust
Estimation of Dispersion Matrices and Principal Components,” Journal
of the American Statistical Association, 76, 354-362.
BIBLIOGRAPHY 481
132. Dixon, W.J., and Tukey, J.W. (1968), “Approximate Behavior of Win-
sorized t (trimming/Winsorization 2),” Technometrics, 10, 83-98.
137. Dollinger, M.B., and Staudte, R.G. (1991), “Influence Functions of It-
eratively Reweighted Least Squares Estimators,” Journal of the Amer-
ican Statistical Association, 86, 709-716.
138. Dongarra, J.J., Moler, C.B., Bunch, J.R., and Stewart, G.W. (1979),
Linpack’s Users Guide, SIAM, Philadelphia, PA.
139. Donoho, D.L., and Huber, P.J. (1983), “The Notion of Breakdown
Point,” in A Festschrift for Erich L. Lehmann, eds. Bickel, P.J., Dok-
sum, K.A., and Hodges, J.L., Wadsworth, Pacific Grove, CA, 157-184.
141. Draper, N.R., and Smith, H. (1981), Applied Regression Analysis, 2nd
ed., John Wiley and Sons, NY.
142. Duda, R.O., Hart, P.E., and Stork, D.G. (2000), Pattern Classification,
2nd ed., John Wiley and Sons, NY.
145. Efron, B., Hastie, T., Johnstone, I. and Tibshirani, R. (2004), “Least
Angle Regression,” (with discussion), The Annals of Statistics, 32, 407-
451.
149. Fan, J., and Li, R. (2001), “Variable Selection via Nonconcave Penal-
ized Likelihood and its Oracle Properties,” Journal of the American
Statistical Association, 96, 1348-1360.
150. Fan, J., and Li, R. (2002), “Variable Selection for Cox’s Proportional
Hazard Model and Frailty Model,” The Annals of Statistics, 30, 74-99.
151. Fang, K.T., and Anderson, T.W. (editors) (1990), Statistical Inference
in Elliptically Contoured and Related Distributions, Allerton Press, NY.
152. Fang, K.T., Kotz, S., and Ng, K.W. (1990), Symmetric Multivariate
and Related Distributions, Chapman & Hall, NY.
153. Farebrother, R.W. (1997), “Notes on the Early History of Elemental Set
Methods,” in L1 -Statistical Procedures and Related Topics, ed. Dodge,
Y., Institute of Mathematical Statistics, Hayward, CA, 161-170.
157. Fowlkes, E.B. (1969), “User’s Manual for a System for Interactive Prob-
ability Plotting on Graphic-2,” Technical Memorandum, AT&T Bell
Laboratories, Murray Hill, NJ.
163. Fung, W.K., He, X., Liu, L., and Shi, P.D. (2002), “Dimension Reduc-
tion Based on Canonical Correlation,” Statistica Sinica, 12, 1093-1114.
164. Furnival, G., and Wilson, R. (1974), “Regression by Leaps and Bounds,”
Technometrics, 16, 499-511.
166. Gather, U., and Becker, C. (1997), “Outlier Identification and Robust
Methods,” in Robust Inference, eds. Maddala, G.S., and Rao, C.R.,
Elsevier Science B.V., Amsterdam, 123-144.
167. Gather, U., Hilker, T., and Becker, C. (2001), “A Robustified Version
of Sliced Inverse Regression,” in Statistics in Genetics and in the Envi-
ronmental Sciences, eds. Fernholtz, T.L., Morgenthaler, S., and Stahel,
W., Birkhäuser, Basel, Switzerland, 145-157.
168. Gather, U., Hilker, T., and Becker, C. (2002), “A Note on Outlier
Sensitivity of Sliced Inverse Regression,” Statistics, 36, 271-281.
173. Golub, G.H., and Van Loan, C.F. (1989), Matrix Computations, 2nd
ed., John Hopkins University Press, Baltimore, MD.
174. Gray, J.B. (1985), “Graphics for Regression Diagnostics,” in the Amer-
ican Statistical Association 1985 Proceedings of the Statistical Comput-
ing Section, 102-108.
175. Greenwood, J.A., and Durand, D. (1960), “Aids for Fitting the Gamma
Distribution by Maximum Likelihood,” Technometrics, 2, 55-56.
179. Hadi, A.S., and Simonoff, J.S. (1993), “Procedures for the Identifica-
tion of Multiple Outliers in Linear Models,” Journal of the American
Statistical Association, 88, 1264-1272.
180. Hahn, G.H., Mason, D.M., and Weiner, D.C. (editors) (1991), Sums,
Trimmed Sums, and Extremes, Birkhäuser, Boston.
181. Hall, P. and Li, K.C. (1993), “On Almost Linearity of Low Dimensional
Projections from High Dimensional Data,” The Annals of Statistics, 21,
867-889.
BIBLIOGRAPHY 485
182. Hall, P., and Welsh, A.H. (1985), “Limit Theorems for the Median
Deviation,” Annals of the Institute of Statistical Mathematics, Part A,
37, 27-36.
183. Hamada, M., and Sitter, R. (2004), “Statistical Research: Some Advice
for Beginners,” The American Statistician, 58, 93-101.
186. Hampel, F.R. (1985), “The Breakdown Points of the Mean Combined
with Some Rejection Rules,” Technometrics, 27, 95-107.
187. Hampel, F.R., Ronchetti, E.M., Rousseeuw, P.J., and Stahel, W.A.
(1986), Robust Statistics, John Wiley and Sons, NY.
188. Hamza, K. (1995), “The Smallest Uniform Upper Bound on the Dis-
tance Between the Mean and the Median of the Binomial and Poisson
Distributions,” Statistics and Probability Letters, 23, 21-25.
189. Härdle, W., Hall, P., and Ichimura, H. (1993), “Optimal Smoothing in
Single Index Models,” The Annals of Statistics, 21, 157-178.
190. Harrison, D. and Rubinfeld, D.L. (1978), “Hedonic Prices and the De-
mand for Clean Air,” Journal of Environmental Economics and Man-
agement, 5, 81-102.
191. Harter, H.L. (1974a), “The Method of Least Squares and Some Alter-
natives, Part I,” International Statistical Review, 42, 147-174.
192. Harter, H.L. (1974b), “The Method of Least Squares and Some Alter-
natives, Part II,” International Statistical Review, 42, 235-165.
193. Harter, H.L. (1975a), “The Method of Least Squares and Some Alter-
natives, Part III,” International Statistical Review, 43, 1-44.
194. Harter, H.L. (1975b), “The Method of Least Squares and Some Alterna-
tives, Part IV,” International Statistical Review, 43, 125-190, 273-278.
BIBLIOGRAPHY 486
195. Harter, H.L. (1975c), “The Method of Least Squares and Some Alter-
natives, Part V,” International Statistical Review, 43, 269-272.
196. Harter, H.L. (1976), “The Method of Least Squares and Some Alter-
natives, Part VI,” International Statistical Review, 44, 113-159.
201. Hawkins, D.M. (1993b), “A Feasible Solution Algorithm for the Min-
imum Volume Ellipsoid Estimator in Multivariate Data,” Computa-
tional Statistics, 9, 95-107.
202. Hawkins, D.M. (1994), “The Feasible Solution Algorithm for the Min-
imum Covariance Determinant Estimator in Multivariate Data, Com-
putational Statistics and Data Analysis, 17, 197-210.
203. Hawkins, D.M., Bradu, D., and Kass, G.V. (1984), “Location of Several
Outliers in Multiple Regression Data Using Elemental Sets,” Techno-
metrics, 26, 197-208.
204. Hawkins, D.M., and Olive, D.J. (1999a), “Improved Feasible Solution
Algorithms for High Breakdown Estimation,” Computational Statistics
and Data Analysis, 30, 1-11.
207. Hawkins, D.M., and Simonoff, J.S. (1993), “High Breakdown Regres-
sion and Multivariate Estimation,” Applied Statistics, 42, 423-432.
209. He, X., Cui, H., and Simpson, D.G. (2004), “Longitudinal Data Anal-
ysis Using t-type Regression,” Journal of Statistical Planning and In-
ference, 122, 253-269.
210. He, X., and Fung, W.K. (1999), “Method of Medians for Lifetime Data
with Weibull Models,” Statistics in Medicine, 18, 1993-2009.
211. He, X., and Fung, W.K. (2000), “High Breakdown Estimation for Mul-
tiple Populations with Applications to Discriminant Analysis,” Journal
of Multivariate Analysis, 72, 151-162.
213. He, X., Simpson, D.G., and Wang, G.Y. (2000), “Breakdown Points of
t-type Regression Estimators,” Biometrika, 87, 675-687.
214. He, X., and Wang, G. (1996), “Cross-Checking Using the Minimum
Volume Ellipsoid Estimator,” Statistica Sinica, 6, 367-374.
215. He, X., and Wang, G. (1997), “A Qualitative Robustness of S*- Estima-
tors of Multivariate Location and Dispersion,” Statistica Neerlandica,
51, 257-268.
221. Hinich, M.J., and Talwar, P.P. (1975), “A Simple Method for Robust
Regression,” Journal of the American Statistical Association, 70, 113-
119.
223. Hoaglin, D.C., Mosteller, F., and Tukey, J.W. (1983), Understanding
Robust and Exploratory Data Analysis, John Wiley and Sons, NY.
224. Hoaglin, D.C., and Welsh, R. (1978), “The Hat Matrix in Regression
and ANOVA,” The American Statistician, 32, 17-22.
225. Horn, P.S. (1983), “Some Easy t-Statistics,” Journal of the American
Statistical Association, 78, 930-936.
231. Hristache, M., Juditsky, A., Polzehl, J., and Spokoiny V. (2001), “Struc-
ture Adaptive Approach for Dimension Reduction,” The Annals of
Statistics, 29, 1537-1566.
232. Huber, P.J. (1981), Robust Statistics, John Wiley and Sons, NY.
235. Iglewicz, B., and Hoaglin, D.C. (1993), How to Detect and Handle Out-
liers, Quality Press, American Society for Quality, Milwaukee, Wiscon-
sin.
236. Insightful (2002), S-Plus 6 Robust Library User’s Guide, Insightful Cor-
poration, Seattle, WA. Available from
(http://math.carleton.ca/∼help/Splus/robust.pdf).
238. Jaeckel, L.A. (1971b), “Some Flexible Estimates of Location,” The An-
nals of Mathematical Statistics, 42, 1540-1552.
242. Johnson, N.L., Kotz, S., and Kemp, A.K. (1992), Univariate Discrete
Distributions, 2nd ed., John Wiley and Sons, NY.
243. Johnson, R.A., and Wichern, D.W. (1988), Applied Multivariate Sta-
tistical Analysis, 2nd ed., Prentice Hall, Englewood Cliffs, NJ.
244. Johnson, R.W. (1996), “Fitting Percentage of Body Fat to Simple Body
Measurements,” Journal of Statistics Education, 4 (1). Available from
(http://www.amstat.org/publications/jse/).
245. Joiner, B.L., and Hall, D.L. (1983), “The Ubiquitous Role of f’/f in
Efficient Estimation of Location,” The American Statistician, 37, 128-
133.
246. Jones, H.L., (1946), “Linear Regression Functions with Neglected Vari-
ables,” Journal of the American Statistical Association, 41, 356-369.
247. Judge, G.G., Griffiths, W.E., Hill, R.C., Lütkepohl, H., and Lee, T.C.
(1985), The Theory and Practice of Econometrics, 2nd ed., John Wiley
and Sons, NY.
248. Jureckova, J., Koenker, R.W., and Welsh, A.H. (1994), “Adaptive
Choice of Trimming Proportions,” Annals of the Institute of Statistical
Mathematics, 46, 737-755.
249. Jureckova, J., and Portnoy, S. (1987), “Asymptotics for One-step M-
estimators in Regression with Application to Combining Efficiency and
High Breakdown Point,” Communications in Statistics Theory and
Methods, 16, 2187-2199.
250. Jureckova, J., and Sen, P.K. (1996), Robust Statistical Procedures:
Asymptotics and Interrelations, John Wiley and Sons, NY.
251. Kafadar, K. (1982), “A Biweight Approach to the One-Sample Prob-
lem,” Journal of the American Statistical Association, 77, 416-424.
252. Kalbfleisch, J.D., and Prentice, R.L. (1980), The Statistical Analysis of
Failure Time Data, John Wiley and Sons, NY.
253. Kay, R., and Little, S. (1987), “Transformations of the Explanatory
Variables in the Logistic Regression Model for Binary Data,” Biometrika,
74, 495-501.
BIBLIOGRAPHY 491
255. Kennedy, W.J., and Gentle, J.E. (1980), Statistical Computing, Marcel
Dekker, NY.
257. Kim, J., and Pollard, D. (1990), “Cube Root Asymptotics,” The Annals
of Statistics, 18, 191-219.
264. Koltchinskii, V.I., and Li, L. (1998), “Testing for Spherical Symmetry
of a Multivariate Distribution,” Journal of Multivariate Analysis, 65,
228-244.
265. Kotz, S., and Johnson, N.L. (editors) (1982ab), Encyclopedia of Statis-
tical Sciences, Vol. 1-2, John Wiley and Sons, NY.
266. Kotz, S., and Johnson, N.L. (editors) (1983ab), Encyclopedia of Statis-
tical Sciences, Vol. 3-4 , John Wiley and Sons, NY.
BIBLIOGRAPHY 492
267. Kotz, S., and Johnson, N.L. (editors) (1985ab), Encyclopedia of Statis-
tical Sciences, Vol. 5-6, John Wiley and Sons, NY.
268. Kotz, S., and Johnson, N.L. (editors) (1986), Encyclopedia of Statistical
Sciences, Vol. 7, John Wiley and Sons, NY.
269. Kotz, S., and Johnson, N.L. (editors) (1988ab), Encyclopedia of Statis-
tical Sciences, Vol. 8-9, John Wiley and Sons, NY.
272. Lax, D.A. (1985), “Robust Estimators of Scale: Finite Sample Perfor-
mance in Long-Tailed Symmetric Distributions,” Journal of the Amer-
ican Statistical Association, 80, 736-741.
273. Le, C.T., (1998), Applied Categorical Data Analysis, John Wiley and
Sons, NY.
275. Lehmann, E.L. (1983), Theory of Point Estimation, John Wiley and
Sons, NY.
277. Li, K.C. (1991), “Sliced Inverse Regression for Dimension Reduction,”
Journal of the American Statistical Association, 86, 316-342.
278. Li, K.C. (1992), “On Principal Hessian Directions for Data Visualiza-
tion and Dimension Reduction: Another Application of Stein’s Lemma,”
Journal of the American Statistical Association, 87, 1025-1040.
280. Li, K.C. (2000), High Dimensional Data Analysis via the SIR/PHD
Approach, Unpublished Manuscript Available from
(http://www.stat.ucla.edu/∼kcli/).
281. Li, K.C., and Duan, N. (1989), “Regression Analysis Under Link Vio-
lation,” The Annals of Statistics, 17, 1009-1052.
282. Li, L., Cook, R.D, and Nachtsheim, C.J. (2004), “Cluster-based Esti-
mation for Sufficient Dimension Reduction,” Computational Statistics
and Data Analysis, 47, 175-193.
283. Li, L., Cook, R.D., and Nachtsheim, C.J. (2005), “Model-Free Variable
Selection,” Journal of the Royal Statistical Society, B, 67, 285-300.
284. Li, R., Fang, K., and Zhu, L. (1997), “Some Q-Q Probability Plots
to Test Spherical and Elliptical Symmetry,” Journal of Computational
and Graphical Statistics, 6, 435-450.
285. Lin, T.C., and Pourahmadi, M. (1998), “Nonparametric and Nonlin-
ear Models and Data Mining in Time Series: A Case-Study on the
Canadian Lynx Data,” Journal of the Royal Statistical Society, C, 47,
187-201.
286. Lindsey, J.K. (2004), Introduction to Applied Statistics: a Modelling
Approach, 2nd ed., Oxford University Press, Oxford, UK.
287. Little, R.J.A., and Rubin, D.B. (2002), Statistical Analysis with Miss-
ing Data, 2nd ed., John Wiley and Sons, NY.
288. Liu, R.Y., Parelius, J.M., and Singh, K. (1999), “Multivariate Analysis
by Data Depth: Descriptive Statistics, Graphics, and Inference,” The
Annals of Statistics, 27, 783-858.
289. Lopuhaä, H.P. (1999), “Asymptotics of Reweighted Estimators of Mul-
tivariate Location and Scatter,” The Annals of Statistics, 27, 1638-
1665.
290. Luo, Z. (1998), “Backfitting in Smoothing Spline Anova,” The Annals
of Statistics, 26, 1733-1759.
291. Ma, Y., and Genton, M.G. (2001), “Highly Robust Estimation of Dis-
persion Matrices,” Journal of Multivariate Analysis, 78, 11-36.
BIBLIOGRAPHY 494
292. Maddela, G.S., and Rao, C.R. (editors) (1997), Robust Inference, Hand-
book of Statistics 15, Elsevier Science B.V., Amsterdam.
293. Maguluri, G., and Singh, K. (1997), “On the Fundamentals of Data
Analysis,” in Robust Inference, eds. Maddela, G.S., and Rao, C.R.,
Elsevier Science B.V., Amsterdam, 537-549.
295. Manzotti, A., Pérez, F.J., and Quiroz, A.J. (2002), “A Statistic for
Testing the Null Hypothesis of Elliptical Symmetry,” Journal of Mul-
tivariate Analysis, 81, 274-285.
298. Mardia, K.V., Kent, J.T., and Bibby, J.M. (1979), Multivariate Anal-
ysis, Academic Press, London.
299. Maronna, R.A., and Zamar, R.H. (2002), “Robust Estimates of Loca-
tion and Dispersion for High-Dimensional Datasets,” Technometrics,
44, 307-317.
300. Maronna, R.A. (2006), Robust Statistics, John Wiley and Sons, NY.
302. MathSoft (1999a), S-Plus 2000 User’s Guide, Data Analysis Products
Division, MathSoft, Seattle, WA. (Mathsoft is now Insightful.)
BIBLIOGRAPHY 495
304. Mayo, M.S., and Gray, J.B. (1997), “Elemental Subsets: the Building
Blocks of Regression,” The American Statistician, 51, 122-129.
305. McCullagh, P., and Nelder, J.A. (1989), Generalized Linear Models,
2nd ed., Chapman & Hall, London.
308. Meeker, W.Q., and Escobar, L.A. (1998), Statistical Methods for Reli-
ability Data, John Wiley and Sons, NY.
310. Moore, D.S. (2004), The Basic Practice of Statistics, 3rd ed., W.H.
Freeman, NY.
311. Moran, P.A.P (1953), “The Statistical Analysis of the Sunspot and
Lynx Cycles,” Journal of Animal Ecology, 18, 115-116.
313. Morgenthaler, S., Ronchetti, E., and Stahel, W.A. (editors) (1993),
New Directions in Statistical Data Analysis and Robustness, Birkhäuser,
Boston.
315. Mosteller, F. (1946), “On Some Useful Inefficient Statistics,” The An-
nals of Mathematical Statistics, 17, 377-408.
BIBLIOGRAPHY 496
316. Mosteller, F., and Tukey, J.W. (1977), Data Analysis and Regression,
Addison-Wesley, Reading, MA.
319. Myers, R.H., Montgomery, D.C., and Vining, G.G. (2002), Generalized
Linear Models with Applications in Engineering and the Sciences, John
Wiley and Sons, NY.
322. Neter, J., Kutner, M.H., Nachtsheim, C.J., and Wasserman W. (1996),
Applied Linear Statistical Models, 4th ed., WcGraw-Hill, Boston, MA.
323. Niinimaa, A., Oja, H., and Tableman, M. (1990), “The Finite-Sample
Breakdown Point of the Oja Bivariate Median and of the Corresponding
Half-Samples Version,” Statistics and Probability Letters, 10, 325-328.
325. Nott, D.J., and Leonte, D. (2004), “Sampling Schemes for Bayesian
Variable Selection in Generalized Linear Models,” Journal of Compu-
tational and Graphical Statistics, 13, 362-382.
326. Olive, D.J. (2001), “High Breakdown Analogs of the Trimmed Mean,”
Statistics and Probability Letters, 51, 87-92.
333. Olive, D.J., and Hawkins, D.M. (2003), “Robust Regression with High
Coverage,” Statistics and Probability Letters, 63, 259-266.
334. Olive, D.J., and Hawkins, D.M. (2005), “Variable Selection for 1D Re-
gression Models,” Technometrics, 47, 43-50.
335. Olive, D.J., and Hawkins, D.M. (2006), “Robustifying Robust Estima-
tors,” Preprint, see (http://www.math.siu.edu/olive/preprints.htm).
338. Patel, J.K., Kapadia C.H., and Owen, D.B. (1976), Handbook of Sta-
tistical Distributions, Marcel Dekker, NY.
339. Peña, D., and Prieto, F.J. (2001), “Multivariate Outlier Detection and
Robust Covariance Matrix Estimation,” Technometrics, 286-299.
BIBLIOGRAPHY 498
341. Pison, G., Rousseeuw, P.J., Filzmoser, P., and Croux, C. (2003), “Ro-
bust Factor Analysis,” Journal of Multivariate Analysis, 84, 145-172.
346. Portnoy, S., and Koenker, R. (1997), “The Gaussian Hare and the
Laplacian Tortoise: Computability of Squared Error Versus Absolute-
Error Estimators,” Statistical Science, 12, 279-300.
348. Poston, W.L., Wegman, E.J., Priebe, C.E., and Solka, J.L. (1997), “A
Deterministic Method for Robust Estimation of Multivariate Location
and Shape,” Journal of Computational and Graphical Statistics, 6, 300-
313.
349. Powers, D.A., and Xie, Y. (2000), Statistical Methods for Categorical
Data Analysis, Academic Press, San Diego.
350. Pratt, J.W. (1959), “On a General Concept of ‘in Probability’,” The
Annals of Mathematical Statistics, 30, 549-558.
BIBLIOGRAPHY 499
353. Price, R.M., and Bonett, D.G. (2001), “Estimating the Variance of the
Sample Median,” Journal of Statistical Computation and Simulation,
68, 295-305.
354. Rao, C.R. (1965), Linear Statistical Inference and Its Applications,
John Wiley and Sons, NY.
356. Rieder, H. (1996), Robust Statistics, Data Analysis, and Computer In-
tensive Methods, Springer-Verlag, NY.
361. Ronchetti, E., and Staudte, R.G. (1994), “A Robust Version of Mal-
lows’s Cp ,” Journal of the American Statistical Association, 89, 550-
559.
BIBLIOGRAPHY 500
365. Rousseeuw, P.J., and Bassett, G.W. (1990), “The Remedian: A Ro-
bust Averaging Method for Large Data Sets,” Journal of the American
Statistical Association, 85, 97-104.
367. Rousseeuw, P.J., and Croux, C. (1992), “Explicit Scale Estimators with
High Breakdown Point,” in L1-Statistical Analysis and Related Meth-
ods, ed. Dodge, Y., Elsevier Science Publishers, Amsterdam, Holland,
77-92.
370. Rousseeuw, P.J., and Leroy, A.M. (1987), Robust Regression and Out-
lier Detection, John Wiley and Sons, NY.
371. Rousseeuw, P.J., Van Aelst, S., Van Driessen, K., and Agulló, J. (2004),
“Robust Multivariate Regression,” Technometrics, 46, 293-305.
372. Rousseeuw, P.J., and Van Driessen, K. (1999), “A Fast Algorithm for
the Minimum Covariance Determinant Estimator,” Technometrics, 41,
212-223.
BIBLIOGRAPHY 501
373. Rousseeuw, P.J., and Van Driessen, K. (2000), “An Algorithm for
Positive-Breakdown Regression Based on Concentration Steps,” in Data
Analysis: Modeling and Practical Application, eds. W. Gaul, O. Opitz,
and M. Schader, Springer-Verlag, NY.
374. Rousseeuw, P.J., and Van Driessen, K. (2002), “Computing LTS Re-
gression for Large Data Sets,” Estadistica, 54, 163-190.
375. Rousseeuw, P.J., and van Zomeren, B.C. (1990), “Unmasking Multi-
variate Outliers and Leverage Points,” Journal of the American Statis-
tical Association, 85, 633-651.
377. Rubin, D.B. (1980), “Composite Points in Weighted Least Squares Re-
gressions,” Technometrics, 22, 343-348.
378. Rubin, D.B. (2004), “On Advice for Beginners in Statistical Research,”
The American Statistician, 58, 196-197.
381. Ruppert, D., and Carroll, R. J. (1980), “Trimmed Least Squares Es-
timation in the Linear Model,” Journal of the American Statistical
Association, 75, 828-838.
382. Satoh, K., and Ohtaki, M. (2004), “A Note on Multiple Regression for
Single Index Model,” Communications in Statistics Theory and Meth-
ods, 33, 2409-2422.
384. Seber, G.A.F., and Lee, A.J. (2003), Linear Regression Analysis, 2nd
ed., John Wiley and Sons, NY.
390. Shorack, G.R., and Wellner, J.A. (1986), Empirical Processes With
Applications to Statistics, John Wiley and Sons, NY.
395. Simonoff, J.S., and Tsai, C. (2002), “Score Tests for the Single Index
Model,” Technometrics, 44, 142-151.
396. Simpson, D.G., Ruppert, D., and Carroll, R.J. (1992), “On One-Step
GM Estimates and Stability of Inferences in Linear Regression,” Jour-
nal of the American Statistical Association, 87, 439-450.
BIBLIOGRAPHY 503
398. Sommer, S., and Huggins, R.M. (1996), “Variables Selection Using the
Wald Test and a Robust Cp ,” Applied Statistics, 45, 15-29.
402. Staudte, R.G., and Sheather, S.J. (1990), Robust Estimation and Test-
ing, John Wiley and Sons, NY.
404. Stefanski, L.A., and Boos, D.D. (2002), “The Calculus of M–estimators,”
The American Statistician, 56, 29-38.
406. Stigler, S.M. (1973b), “Simon Newcomb, Percy Daniell, and the History
of Robust Estimation 1885-1920,” Journal of the American Statistical
Association, 68, 872-878.
407. Stigler, S.M (1977), “Do Robust Estimators Work with Real Data?”
The Annals of Statistics, 5, 1055-1098.
409. Street, J.O., Carroll, R.J., and Ruppert, D. (1988), “A Note on Com-
puting Regression Estimates Via Iteratively Reweighted Least Squares,”
The American Statistician, 42, 152-154.
BIBLIOGRAPHY 504
412. Stromberg, A.J., Hawkins, D.M., and Hössjer, O. (2000), “The Least
Trimmed Differences Regression Estimator and Alternatives,” Journal
of the American Statistical Association, 95, 853-864.
413. Tableman, M. (1994a), “The Influence Functions for the Least Trimmed
Squares and the Least Trimmed Absolute Deviations Estimators,” Statis-
tics and Probability Letters, 19, 329-337.
415. Thode, H.C. (2002), Testing for Normality, Marcel Dekker, NY.
418. Tong, H. (1977), “Some Comments on the Canadian Lynx Data,” Jour-
nal of the Royal Statistical Society, A, 140, 432-468.
425. Tukey, J.W., and McLaughlin, D.H. (1963), “Less Vulnerable Con-
fidence and Significance Procedures for Location Based on a Single
Sample: Trimming/Winsorization 1,” Sankhya, A, 25, 331-352.
428. Velleman, P.F., and Welsch, R.E. (1981), “Efficient Computing of Re-
gression Diagnostics,” The American Statistician, 35, 234-242.
429. Venables, W.N., and Ripley, V.D. (1997), Modern Applied Statistics
with S-PLUS, 2nd ed., Springer-Verlag, NY.
430. Vı́šek, J.Á. (1996), “On High Breakdown Point Estimation,” Compu-
tational Statistics, 11, 137-146.
431. Wackerly, D.D., Mendenhall, W., and Scheaffer, R.L., (2002), Mathe-
matical Statistics with Applications, 6th ed., Duxbury, Pacific Grove,
CA.
432. Wand, M.P. (1999), “A Central Limit Theorem for Local Polynomial
Backfitting Estimators,” Journal of Multivariate Analysis, 70, 57-65.
434. Weisberg, S. (2005), Applied Linear Regression, 3rd ed., John Wiley
and Sons, NY.
435. Weisberg, S., and Welsh, A.H. (1994), “Adapting for the Missing Link,”
The Annals of Statistics, 22, 1674-1700.
BIBLIOGRAPHY 506
436. Welch, B.L. (1937), “The Significance of the Difference Between Two
Means When the Population Variances are Unequal,” Biometrika, 29,
350-362.
437. Welsh, A.H. (1986), “Bahadur Representations for Robust Scale Esti-
mators Based on Regression Residuals,” The Annals of Statistics, 14,
1246-1251.
443. Willems, G., Pison, G., Rousseeuw, P.J., and Van Aelst, S. (2002), “A
Robust Hotelling Test,” Metrika, 55, 125-138.
446. Woodruff, D.L., and Rocke, D.M. (1993), “Heuristic Search Algorithms
for the Minimum Volume Ellipsoid,” Journal of Computational and
Graphical Statistics, 2, 69-95.
447. Woodruff, D.L., and Rocke, D.M. (1994), “Computable Robust Esti-
mation of Multivariate Location and Shape in High Dimension Using
Compound Estimators,” Journal of the American Statistical Associa-
tion, 89, 888-896.
BIBLIOGRAPHY 507
448. Xia, Y., Tong, H., Li, W.K., and Zhu, L.-X. (2002), “An Adaptive
Estimation of Dimension Reduction Space,” (with discussion), Journal
of the Royal Statistical Society, B, 64, 363-410.
449. Yeo, I.K., and Johnson, R. (2000), “A New Family of Power Transfor-
mations to Improve Normality or Symmetry,” Biometrika, 87, 954-959.
450. Yin, X.R., and Cook, R.D. (2002), “Dimension Reduction for the Con-
ditional kth Moment in Regression,” Journal of the Royal Statistical
Society, B, 64, 159-175.
451. Yin, X., and Cook, R.D. (2003), “Estimating Central Subspaces Via
Inverse Third Moments,” Biometrika, 90, 113-125.
453. Yuen, K.K. (1974), “The Two-Sample Trimmed t for Unequal Popula-
tion Variances,” Biometrika, 61, 165-170.
508
INDEX 509
Welch, 43
Welch intervals, 43
Wellner, 51, 52, 64, 216
Welsch, vii, 202
Welsh, 63, 64, 202, 203, 217, 228,
248, 373
White, ix, 303
Whitten, 89, 92
Wichern, ix, 286, 296, 300, 311
Wilcox, vii, 64
Wilcoxon rank estimator, 228, 342
Wilks, 19, 441
Willems, 332
Wilson, 143, 166, 360
Wilson–Hilferty approximation, 76,
82
Winkelmann, 417
Winsor’s principle, 347
Winsorized mean, 44, 63
Winsorized random variable, 50, 102
Wishart distribution, 249
Wisnowski, 312
Wood, 143, 362
Woodruff, 264, 272, 297, 310, 311
Wu, 166
Zamar, 310
Zhu, 319, 332, 372, 438
Zuo, 275, 295