Ch07-Point Estimation of Parameters and Sampling Distributions
Ch07-Point Estimation of Parameters and Sampling Distributions
for Engineers
Seventh Edition
Douglas C. Montgomery George C. Runger
Chapter 7
Point Estimation of Parameters and Sampling Distributions
Chapter 7 Title Slide
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved 1
7 Point Estimation of Parameters and
Sampling Distributions
CHAPTER OUTLINE
7.1 Point Estimation 7.3.4 Bootstrap Standard Error
7.2 Sampling Distributions and 7.3.5 Mean Squared Error of an
the Central Limit Theorem Estimator
7.3 General Concepts of Point 7.4 Methods of Point Estimation
Estimation 7.4.1 Method of Moments
7.3.1 Unbiased Estimators 7.4.2 Method of Maximum
7.3.2 Variance of a Point Likelihood
Estimator 7.4.3 Bayesian Estimation of
7.3.3 Standard Error: Reporting Parameters
a Point Estimate
Chapter 7 Contents
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
2
Learning Objectives for Chapter 7
After careful study of this chapter, you should be able to do the following:
1. Explain the general concepts of estimating the parameters of a population or a
probability distribution
2. Explain the important role of the normal distribution as a sampling distribution and
the central limit theorem
3. Explain important properties of point estimators, including bias, variances, and
mean square error
4. Construct point estimators using the method of moments and the method of
maximum likelihood.
5. Compute and explain the precision with which a parameter is estimated
6. Construct a point estimator using the Bayesian approach
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
3
Point Estimation
• A point estimate is a reasonable value of a population
parameter.
• 𝑋1, 𝑋2, … , 𝑋𝑛 are random variables.
• Functions of these random variables, 𝑥ҧ and 𝑠2, are also random
variables called statistics.
• Statistics have their unique distributions which are called
sampling distributions.
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
4
Point Estimator
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
5
Some Parameters & Their Statistics
• There could be choices for the point
estimator of a parameter.
• To estimate the mean of a population,
we could choose the:
▪ Sample mean.
▪ Sample median.
▪ Average of the largest & smallest
observations in the sample.
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
6
Some Definitions
• The random variables 𝑋1, 𝑋2, … , 𝑋𝑛 are a random sample of
size 𝑛 if:
a) The 𝑋𝑖 ′𝑠 are independent random variables.
b) Every 𝑋𝑖 has the same probability distribution.
• A statistic is any function of the observations in a random
sample.
• The probability distribution of a statistic is called a sampling
distribution.
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
7
Central Limit Theorem
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
8
Example 7.2 | Central Limit Theorem
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
9
Sampling Distribution of a Difference in
Sample Means
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
10
Unbiased Estimators
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
11
Example 7.3 | Sample Mean and Variance
are Unbiased
• Suppose 𝑋 is a random variable with mean 𝜇 and variance 𝜎2. Let 𝑋1, 𝑋2, … , 𝑋𝑛 be a
random sample of size 𝑛 from the population represented by 𝑋
• Show that the sample mean 𝑋ത and sample variance 𝑆 2 are unbiased estimators of 𝜇 and
𝜎 2 respectively.
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
12
Variance of a Point Estimator
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
13
Standard Error: Reporting a Point Estimate
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
14
Example 7.4 | Thermal Conductivity
• The following 10 measurements of thermal conductivity of Armco iron were obtained:
• A point estimate of the mean thermal conductivity at 100 ℉ and 550 watts is the sample
mean or
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
16
Mean Squared Error of an Estimator
The mean squared error is equal to the variance of the estimator plus the squared bias
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
17
Mean Squared Error of an Estimator
If this relative efficiency is less than 1, we
would conclude that the first estimator is a
more efficient estimator than the second
estimator, in the sense that it has a smaller
mean squared error
An estimator that has a mean squared error that is less than or equal to the mean
squared error of any other estimator, for all values of the parameter estimator, is called
an optimal estimator of Θ. Optimal estimators rarely exist.
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
18
Methods of Point Estimation
• The general idea behind the method of moments is to equate population moments,
which are defined in terms of expected values, to the corresponding sample moments
n i =1 n i =1
2
1 n
n
1 n 2
X i
2
− n n iX
i =1
= X i − X 2 = i =1
2
n i =1 n
n
2
n
Xi ( Xi − X )
2
1 n 2 i =1
= Xi − = i =1
(biased)
n i =1 n n
Sec 7.4.1 Methods of Point Estimation
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
20
Example 7.8 | Gamma Distribution Moment
Estimators
Suppose that 𝑋1, 𝑋2, … , 𝑋𝑛 is a random sample from a gamma distribution with parameter r
and λ where 𝐸(𝑋) = 𝑟/ 𝜆 and 𝐸(𝑋2) = 𝑟(𝑟 + 1)/ 𝜆2 .
X2 21.6462
r= = = 1.29
n
(1/ n ) X i2 − X 2 ( )
1 8 6645.4247 − 21.646 2
i =1
X 21.646
= = = 0.0598
n
(1/ n ) X i2 − X 2 (1 8 ) 6645.4247 − 21.6462
i =1
n
n xi n
xi
=p (1 − p ) (1 − p )
1− xi n−
xi
=p i =1
i =1
i =1
n n
ln L ( p ) = xi ln p + n − xi ln (1 − p )
i =1 i =1
n
n
d ln L ( p )
xi
n − i =1
xi
= i =1
−
dp p (1 − p )
Sec 7.4.2 Method of Maximum Likelihood
Copyright © 2019 John Wiley & Sons, Inc. All Rights Reserved
23
Example 7.11 | Normal Distribution for MLEs
for 𝜇 and 𝜎2
Let 𝑋 be a normal random variable with unknown mean 𝜇 and known variance 𝜎2. The
likelihood function of a random sample of size 𝑛 is:
Conclusion: Once again, the maximum likelihood estimators are equal to the moment estimators
1. It may not be easy to maximize the likelihood function because the derivative
function set to zero may be difficult to solve algebraically.
2. It may not always be possible to use calculus methods directly to determine the
maximum of 𝐿(𝜃).
f ( x1 , x2 ,..., xn ,θ )
• The desired posterior distribution is f ( θ | x1 , x2 ,..., xn ) =
f ( x1 , x2 ,..., xn )
( ) ( x − )
n
− 1 2 2 2
1
f ( x1 , x2 ,..., xn | ) =
i
i =1
e
( 2 )2 n 2
n 2
( )
n
1 − 1 2 2
xi − 2 xi + n 2
= e i =1 i =1
( 2 )2 n 2