Chapter6 01
Chapter6 01
Point Estimation
6.1
10
Example 2
Reconsider the accompanying 20 observations on dielectric
breakdown voltage for pieces of epoxy resin.
24.46
27.98
25.61
28.04
26.25
28.28
26.42
28.49
27.94
30.88
Example 2
contd
12
Example 2
contd
13
Example 2
contd
14
15
16
17
Unbiased Estimators
18
Unbiased Estimators
Suppose we have two measuring instruments; one
instrument has been accurately calibrated, but the other
systematically gives readings smaller than the true value
being measured.
When each instrument is used repeatedly on the same
object, because of measurement error, the observed
measurements will not be identical.
However, the measurements produced by the first
instrument will be distributed about the true value in such a
way that on average this instrument measures what it
purports to measure, so it is called an unbiased instrument.
19
Unbiased Estimators
The second instrument yields observations that have a
systematic error component or bias.
Definition
A point estimator is said to be an unbiased estimator of
if E( ) = for every possible value of . If is not
unbiased, the difference E( ) is called the bias of .
That is, is unbiased if its probability (i.e., sampling)
distribution is always centered at the true value of the
parameter.
20
Unbiased Estimators
Suppose is an unbiased estimator; then if = 100, the
sampling distribution is centered at 100; if = 27.5, then
the sampling distribution is centered at 27.5, and so on.
Figure 6.1 pictures the distributions of several biased and
unbiased estimators. Note that centered here means that
the expected value, not the median, of the distribution of
is equal to .
for a parameter
21
Unbiased Estimators
It may seem as though it is necessary to know the value of
(in which case estimation is unnecessary) to see whether
is unbiased.
This is not usually the case, though, because unbiasedness
is a general property of the estimators sampling
distributionwhere it is centeredwhich is typically not
dependent on any particular parameter value.
The sample proportion X/n can be used as an estimator of
p, where X, the number of sample successes, has a
binomial distribution with parameters n and p.
22
Unbiased Estimators
Thus
E( ) = E
E(X) =
(np) = p
Proposition
When X is a binomial rv with parameters n and p, the
sample proportion = X/n is an unbiased estimator of p.
No matter what the true value of p is, the distribution of the
estimator will be centered at the true value.
23
Example 4
Suppose that X, the reaction time to a certain stimulus, has
a uniform distribution on the interval from 0 to an unknown
upper limit (so the density function of X is rectangular in
shape with height 1/ for 0 x ).
It is desired to estimate on the basis of a random sample
X1, X2, . . . , Xn of reaction times.
Since is the largest possible time in the entire population
of reaction times, consider as a first estimator the largest
sample reaction time: = max (X1 ,..., Xn).
24
Example 4
contd
25
Example 4
contd
max(X1 ,, Xn)
26
Example 4
contd
max(X1 ,, Xn)
27
Unbiased Estimators
Principle of Unbiased Estimation
When choosing among several different estimators of ,
select one that is unbiased.
According to this principle, the unbiased estimator in
Example 4 should be preferred to the biased estimator .
Consider now the problem of estimating 2.
28
Unbiased Estimators
Proposition
Let X1, X2, . . . , Xn be a random sample from a distribution
with mean and variance 2. Then the estimator
29
Unbiased Estimators
This estimator is therefore not unbiased. The bias is
(n 1) 2/n 2 = 2 /n.
Because the bias is negative, the estimator with divisor n
tends to underestimate 2, and this is why the divisor n 1
is preferred by many statisticians (though when n is large,
the bias is small and there is little difference between the
two).
Unfortunately, the fact that S2 is unbiased for estimating 2
does not imply that S is unbiased for estimating .
30
Unbiased Estimators
Taking the square root messes up the property of
unbiasedness (the expected value of the square root is not
the square root of the expected value).
Fortunately, the bias of S is small unless n is quite small.
There are other good reasons to use S as an estimator,
especially when the population distribution is normal.
31
Unbiased Estimators
In Example 2, we proposed several different estimators for
the mean of a normal distribution.
If there were a unique unbiased estimator for , the
estimation problem would be resolved by using that
estimator. Unfortunately, this is not the case.
Proposition
If X1, X2, . . . , Xn is a random sample from a distribution
with mean , then X is an unbiased estimator of . If in
addition the distribution is continuous and symmetric,
then and any trimmed mean are also unbiased estimators
of .
32
Unbiased Estimators
The fact that X is unbiased is just a restatement of one of
our rules of expected value: E(X ) = for every possible
value of (for discrete as well as continuous distributions).
The unbiasedness of the other estimators is more difficult
to verify.
The next example introduces another situation in which
there are several unbiased estimators for a particular
parameter.
33
Example 5
Under certain circumstances organic contaminants adhere
readily to wafer surfaces and cause deterioration in
semiconductor manufacturing devices.
The paper Ceramic Chemical Filter for Removal of
Organic Contaminants (J. of the Institute of Environmental
Sciences and Technology, 2003: 5965) discussed a
recently developed alternative to conventional charcoal
filters for removing organic airborne molecular
contamination in cleanroom applications.
34
Example 5
contd
2
1.3
1.1
3
1.5
4.5
4
3.0
3.5
5
11.6
14.4
6
26.6
29.1
35
Example 5
contd
36
Example 5
contd
Example 5
contd
Example 5
contd
But all three are unbiased; the argument relies on the fact
that each one is a linear function of the Yis (we are
assuming here that the xis are fixed, not random):
39
40
41
42
43
Example 6
We argued in Example 4 that when X1, . . . , Xn is a random
sample from a uniform distribution on [0, ], the estimator
Example 6
contd
45
Some Complications
48
Some Complications
The last theorem does not say that in estimating a
population mean , the estimator X should be used
irrespective of the distribution being sampled.
49
Example 7
Suppose we wish to estimate the thermal conductivity of a
certain material. Using standard measurement techniques,
we will obtain a random sample X1, . . . , Xn of n thermal
conductivity measurements.
Lets assume that the population distribution is a member of
one of the following three families:
(6.1)
(6.2)
50
Example 7
contd
(6.3)
51
Example 7
contd
Example 7
contd
Example 7
contd
54
55
56
57
Example 9
Example 2 continued
Assuming that breakdown voltage is normally distributed,
is the best estimator of . If the value of is known to
be 1.5, the standard error of X
is
.
If, as is usually the case, the value of is unknown, the
estimate = s = 1.462 is substituted into
to obtain the
estimated standard error
.
58
59
61
; estimate =
; estimate =
; estimate =
, the
62
63
Example 11
A theoretical model suggests that X, the time to breakdown
of an insulating fluid between electrodes at a particular
voltage, has f (x; ) = ex, an exponential distribution.
A random sample of n = 10 breakdown times (min) gives
the following data:
41.53 18.73 2.99 30.34 12.33 117.52 73.02 223.63 4.00 26.78
Example 11
contd
The first such sample was 41.00, 109.70, 16.78, 6.31, 6.76,
5.62, 60.96, 78.81, 192.25, 27.61, from which
= 545.8
and
= 1/ 54.58 = .01832.
The average of the 100 bootstrap estimates is = .02153,
and the sample standard deviation of these 100 estimates
is = .0091, the bootstrap estimate of s standard error.
A histogram of the 100 s was somewhat positively
skewed, suggesting that the sampling distribution of also
has this property.
65