0% found this document useful (0 votes)
37 views

Unit - III (P&S Notes)

This document discusses sampling and parameter estimation. It covers topics like sample mean, central limit theorem, sampling distributions, maximum likelihood estimators, interval estimates using chi-square, t and F distributions. It also discusses different types of sampling, properties of estimators, and point and interval estimation.

Uploaded by

ranjithmulugu1
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
37 views

Unit - III (P&S Notes)

This document discusses sampling and parameter estimation. It covers topics like sample mean, central limit theorem, sampling distributions, maximum likelihood estimators, interval estimates using chi-square, t and F distributions. It also discusses different types of sampling, properties of estimators, and point and interval estimation.

Uploaded by

ranjithmulugu1
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 39

Unit – III

Sampling and Parameter Estimation


Introduction-Sample mean-Central Limit theorem,
sample variance-sampling distribution from a normal
population-sampling from a finite population,
Maximum likelihood estimators – Interval Estimates
–Chi-Square distribution
– t-distribution
– F-distribution.
Population & Samples
Population is the totality of the observation with which we
are consider. The number of observations in the
population is defined to be the “size of the population”

➢ If the size is finite, the population is said to be “finite


population”
➢ If the size is infinite, the population is said to be
“infinite population”

A sample is a subset of a population. The number of


objects in the sample is called the “size of the sample”. The
statistical constants like mean, SD, Correlation coefficient
etc., are obtained for the population are called
“parameters”
Types of sampling :

1) Purposive sampling
2) Random Sampling
3) Simple sampling
4) Stratified sampling

Purposive sampling:
A purposive sampling is one which the samples in it are selected with
definite purpose in view

Random sampling:
It is one which each member of the population has an equal of being
included in the sample
Simple Sampling

It is a random sampling in which each member of population has an


equal chances of being included in the sample and that this probability
is independent of the previous drawing.

Stratified Sampling

The sample which is aggregate of sample units of each of the stratum


(society of group of people where similar in education) is called is
called Stratified Sampling
Observation:
In this chapter, we will be concerned with the probability distributions
of certain statistics that arise from a sample, where a statistic is a
random variable whose value is determined by the sample data.

Two important statistics that we will discuss are:


Sample mean and the sample variance.
Here, we consider the sample mean and derive its expectation and
variance.
We note that when the sample size is at least moderately large, the
distribution of the sample mean is approximately normal. This follows
from the central limit theorem, one of the most important theoretical
results in probability.
Classification of Samples
➢ Large Sample
➢ Small Sample

Large Sample :
If the size of the sample 𝑛 ≥ 30, the sample is said to be large sample
Small Sample :
If the size of the sample 𝑛 < 30, the sample is said to be small sample
or exact sample.

Formulae:
𝜇 − 𝑃𝑜𝑝𝑢𝑙𝑎𝑡𝑖𝑜𝑛 𝑚𝑒𝑎𝑛
𝑛
2
2
𝑋𝑖 − 𝜇
𝜎 − 𝑃𝑜𝑝𝑢𝑙𝑎𝑡𝑖𝑜𝑛 𝑣𝑎𝑟𝑖𝑒𝑛𝑐𝑒 = ෍
𝑁
𝑖=1
𝑛
𝑋𝑖
𝑋ത − 𝑆𝑎𝑚𝑝𝑙𝑒 𝑚𝑒𝑎𝑛 = 𝜇𝑋ത = ෍
𝑁
𝑖=1
𝑛
2 2 𝑋𝑖 − 𝑋ത 2
𝑆 − 𝑆𝑎𝑚𝑝𝑙𝑒𝑣𝑎𝑟𝑖𝑒𝑛𝑐𝑒 = 𝜎𝑋ത = ෍
𝑁−1
𝑖=1
𝜎
𝑆𝑎𝑚𝑝𝑙𝑒 𝑠𝑖𝑧𝑒 =
𝑛

Expected value and variance are obtained as follows:


Expectation:
𝑋1 + 𝑋2 + ⋯ 𝑋𝑛
𝐸 𝑋ത = 𝐸
𝑛
1
= 𝐸 𝑋1 + 𝐸 𝑋2 + ⋯ 𝐸 𝑋𝑛
𝑛
=𝜇
Variance:
𝑿𝟏 + 𝑿𝟐 + ⋯ 𝑿𝒏
ഥ =𝑽
𝑽𝑿
𝒏
𝟏
= 𝑽 𝑿𝟏 + 𝑽 𝑿𝟐 + ⋯ 𝑽 𝑿𝒏
𝒏𝟐
𝒏𝝈𝟐 𝝈𝟐
= 𝟐 =
𝒏 𝒏

The central limit theorem


Let 𝑋1 , 𝑋2 , 𝑋3 …. 𝑋𝑛 be a sequence of independent and identically
distributed random variables each having mean 𝜇 and variance 𝜎 2 ,
Then for n large, the distribution of 𝑋1 + 𝑋2 + 𝑋3 + ⋯ 𝑋𝑛 is approximately
normal with mean 𝑛𝜇 and variance 𝑛𝜎 2
𝑋1 +𝑋2 +𝑋3 +⋯𝑋𝑛 −𝑛𝜇
Therefore,
𝜎 𝑛
is approximately a standard normal random variable; thus, for n large,
𝑋1 + 𝑋2 + 𝑋3 + ⋯ 𝑋𝑛 − 𝑛𝜇
𝑃 <𝑥 ≈𝑃 𝑍<𝑥
𝜎 𝑛
Example: An insurance company has 25,000 automobile policy holders.
If the yearly claim of a policy holder is a random variable with mean
320 and standard deviation 540, approximate the probability that the
total yearly claim exceeds 8.3 million.
Solution:
Let 𝑋 denote the total yearly claim. Number the policy holders, and let
𝑋𝑖 denote the yearly claim of policy holder𝑖. With n = 25,000, we have
from the central limit theorem that
𝑋 = σ𝑛𝑖=1 𝑋𝑖 will have approximately a normal distribution with
Mean=320 × 25000 = 8 × 106
and Standard deviation= 540 25,000 = 8.5381 × 104

Therefore,
𝑋 − 8 × 106 8.3 × 106 − 8 × 106
𝑃 𝑋 > 8.3 × 106 =𝑃 4
>
8.5381 × 10 8.5381 × 104
𝑋 − 8 × 106 0.3 × 106
=𝑃 4
>
8.5381 × 10 8.5381 × 104
= 𝑃 𝑍 > 3.51 , 𝑤ℎ𝑒𝑟𝑒 𝑍 𝑖𝑠 𝑠𝑡𝑎𝑛𝑑𝑎𝑟𝑑 𝑛𝑜𝑟𝑚𝑎𝑙
≅ 0.00023

Thus, there are only 2.3 chances out of 10,000 that the total yearly
claim will exceed 8.3 million

Example 2:
The ideal size of a first-year class at a particular college is 150
students. The college, knowing from past experience that, on the
average, only 30 percent of those accepted for admission will actually
attend, uses a policy of approving the applications of 450 students.
Compute the probability that more than 150 first-year students attend
this college.
Let X denote the number of students that attend; then assuming that
each accepted applicant will independently attend, it follows that X is
a binomial random variable with parameters n = 450 and p = 0.3.
Since the binomial is a discrete and the normal a continuous
distribution, it is best to compute

𝑃 𝑋 = 𝑖 𝑎𝑠 𝑃 𝑖 − 0.5 < 𝑋 < 𝑖 + 0.5

when applying the normal approximation, this yields the


approximation
(X − 450 (0.3) ) (150.5 − 450 (0.3) )
𝑃 𝑋 > 150.5 = 𝑃 >
450(0.3)(0.5) 450(0.3)(0.5)
𝑃 𝑍 > 1.59 = 0.06
Hence, only 6 percent of the time do more than 150 of the first 450
accepted actually attend.
Sampling distributions from a normal population

Let 𝑋1 , 𝑋2 , 𝑋3, … 𝑋𝑛 be a sample from a normal population having


mean 𝜇 and variance 𝜎 2 . That is, they are independent and
𝑋𝑖 ~𝑁(𝜇, 𝜎 2 ), 𝑖 = 1,2, … 𝑛. Also let
𝑛
𝑋𝑖
𝑋ത = ෍
𝑛
𝑖=1

σ𝑛 𝑋 −𝑋ത 2
𝑠2 = 𝑖=1 𝑖
𝑛−1
How to compute distribution of sample mean and sample variance
1) Distribution of Sample mean
Since the sum of independent normal random variables is normally
distributed, it follows that 𝑋ത is normal with mean
𝑛
𝐸 𝑋𝑖

𝐸 𝑋 =෍ =𝜇
𝑛
𝑖=1

2) Distribution of Sample variance


1 𝑛 𝜎2
𝑉𝑎𝑟 𝑋ത = σ𝑖=1 𝑉𝑎𝑟 𝑋𝑖 =
𝑛 𝑛
Note:
ത the average of the sample, is normal with a mean equal to
That is, 𝑋,
the population mean but with a variance reduced by a factor of 1/𝑛.
It follows from this that

𝑋−𝜇
is a standard normal variable.
𝜎/ 𝑛
Example 1: A population consists of 5 no’s 2,3,6,8,11. Consider all
possible size 2 that can be drawn with replacement from the
population. Find
(i) Mean of population
(ii) S.D of population
(iii) Mean of sample distribution of means
(iv) S.D. of sample distribution of means

Example 2: If a population is 3,6,9,15,27. Consider all possible size 3


that can be drawn without replacement from the population. Find
(i) Mean of population
(ii) S.D of population
(iii) Mean of sample distribution of means
(iv) S.D. of sample distribution of means
Estimation

Estimation

Estimate

Unknown population 𝒑, 𝝀, 𝝁, 𝝈
Parameters
Estimators

Sampling Population
Mean (𝑿
ഥ) Mean (𝛍)

Types of Estimators

Point estimator Interval estimator Bayesian estimator


Properties of estimators
(i) Unbiased estimator
(ii) More efficient estimator
(iii) Most efficient estimator

Estimate:
An estimate is a statement made to find an unknown population
parameters quantities appearing in distribution such as ‘p’ in
binomial, 𝝁, 𝝈 in normal distribution.

Estimator:
The procedure or rule to determine an unknown population
parameter is called an estimator.
Example : Sample mean 𝑿 ഥ is an estimator of population mean 𝝁
because sample mean is a method of determining population
mean. A parameter can have one or two or many estimators.
The estimation can be done in two ways
(i) Point estimation
(ii) Interval estimation

Statistical estimation:
It is a part of statistical inference where a population parameter is
estimator from the corresponding sample statistics.

Point Estimation:
A point estimation of a parameter is a statistical estimation where
the parameter is estimated by a single numerical value from sample
data.
❖ A point estimate of parameter 𝜽 is a single numerical value
which is computed from a given sample has an approximation of
the unknown exact value of the parameter.
❖ A point estimator is a static for estimating the population
parameter 𝜽 and it is denoted by 𝜽෡
❖ The value 𝒙 ഥ computed from a sample size ‘n’ is
ഥ of the statistics 𝑿

a point estimate of the population parameter 𝛍 i.e., 𝝁 = 𝑿
Properties of estimation
An estimator is not expected to estimate the population parameter
without error. An estimator should be close to the true value for
unknown parameter.
(a) Unbiased estimator
෡ be an estimator of 𝜽, the static 𝜽 is said to be unbiased estimator
𝜽
or its value in an unbiased estimate if and only if the mean
expected value of 𝜽 ෡ = 𝜽 . i.e. the mean of probability distribution
෡ is equal to 𝜽.
of 𝜽

• An estimator possessing this property is said to be unbiased.


෡ is said to be an unbiased estimator
• A static or point estimator 𝜽
෡ = 𝜽. i.e. E (Statistic)=parameter,
of the parameter 𝜽, if 𝑬 𝜽
• Then static is said to be unbiased estimator of the parameter.

Eg: 𝑺𝟐 is an unbiased estimator of the parameter

Variance of point estimator:


෡ 𝟏 and 𝜽
If 𝜽 ෡ 𝟐 are two unbiased estimators of the population
parameter 𝜽, we would choose the estimator whose sampling
distribution has the smaller variance. Hence 𝝈𝟐𝜽෡𝟏 < 𝝈𝟐𝜽෡𝟐 . Hence 𝜽
෡ 𝟏 is
more efficient estimator of 𝜽, then comes 𝜽෡𝟐.

Most efficient estimator


The one with the smallest variance is called the most efficient
estimator of 𝜽.
Good estimator:
An estimator is said to be good if it is (i) unbiased (ii) consistent (iii)
efficient.
Interval estimation:
Point estimation really coincide with quantities. These are
interested to estimate, so instead of point estimation we can use
the interval estimations.
▪ In point estimation we can take single numerical value but in
interval estimations, we can take interval values so interval
estimation is the better way to estimate.

Interval Estimate:
Even the most efficient unbiased estimator cannot estimate the
population parameter exactly. It is true, that our ocuurency
increased with large sample.

In many situations, it is preferable to determine interval within


which we would accept to find the value of the parameter. Such an
interval is called an interval estimate.
Note: An interval estimate is an interval obtained from a sample.
Interval Estimation definition:
An interval estimate of a population parameter 𝜽is an interval of
the form 𝜽 ෡𝑳 < 𝜽 < 𝜽 ෡ 𝒔 , where 𝜽
෡ 𝑳 and 𝜽
෡ 𝒔 are end points of the
interval of values of corresponding random variables
𝑷 𝜽 ෡𝑳 < 𝜽 < 𝜽෡𝒔 = 𝟏 − 𝜶

Where 𝟏 − 𝜶 is confidence coefficient

• If we take 𝜶 values between 0 &1 then end points are taken


from an unknown sample distribution.

• 𝟏 − 𝜶 100% is called confidence interval

• If 𝜶 = 𝟎. 𝟎𝟓 we get 95%

• If 𝜶 = 𝟎. 𝟎1 we get 99%
Maximum error of estimate E for large sample
Since the sample mean estimate very rarely is equal to the mean
of population 𝝁 , a point estimate is generally accompined with a
statement of error which gives difference b/w estimate and the
quality to be estimated.

ഥ−𝝁
Therefore, estimator error is 𝒙
ഥ−𝝁
𝒙
For large n, the random variable is normal variate
𝝈/ 𝒏
approximately then

ഥ−𝝁
𝒙
𝑷 −𝒛𝜶/𝟐 < 𝒛 < 𝒛𝜶/𝟐 = 𝟏 − 𝜶 where 𝐳 =
𝝈/ 𝒏

ഥ−𝝁
𝒙
Hence 𝑷 −𝒛𝜶/𝟐 < < 𝒛𝜶/𝟐 = 𝟏 − 𝜶
𝝈/ 𝒏
𝑨𝒄𝒄𝒆𝒑𝒕𝒂𝒃𝒍𝒆 𝒓𝒆𝒈𝒊𝒐𝒏

𝑹𝒆𝒋𝒆𝒄𝒕𝒊𝒐𝒏 𝒓𝒆𝒈𝒊𝒐𝒏 𝑹𝒆𝒋𝒆𝒄𝒕𝒊𝒐𝒏 𝒓𝒆𝒈𝒊𝒐𝒏

−𝒛𝜶/𝟐 𝒛=𝟎 𝒛𝜶/𝟐

ഥ from each term


Multiplying each term in the inequality by 𝝈/ 𝒏 and subtracting 𝒙
multiplying by ‘-1’
ഥ−𝝁
𝒙
ഥ − 𝒛𝜶/𝟐 𝝈/ 𝒏 <
𝑷 𝒙 ഥ + 𝒛𝜶/𝟐 𝝈/ 𝒏
<𝒙
𝝈/ 𝒏
Where 𝒙ഥ is the mean of the random sample of size ‘n’ from the
population with known variance 𝝈𝟐
𝒛𝜶/𝟐 is the 𝒛 value leaving an area 𝜶/𝟐 to the right side. So,
Maximum error of estimate E with (𝟏 − 𝜶) probability is given by
𝝈
𝑬 = 𝒛𝜶/𝟐
𝒏
Where 𝜶, 𝑬, 𝝈 are known, then sample size ‘n’ is given by

𝝈 𝒛𝜶/𝟐 𝟐
𝒏=
𝑬
Maximum error of estimate E for small sample
When 𝒏 < 𝟑𝟎, small sample , we use s, the standard deviation of
sample to determine 𝑬. When 𝝈 is known , 𝒕 can be used to
construct confidence interval as 𝝁
Hence by the previous process

ഥ−𝝁
𝒙
𝑷 −𝒕𝜶/𝟐 < 𝒕 < 𝒕𝜶/𝟐 = 𝟏 − 𝜶 where 𝒕 =
𝒔/ 𝒏

ഥ−𝝁
𝒙
Hence 𝑷 −𝒕𝜶/𝟐 < < 𝒕𝜶/𝟐 = 𝟏 − 𝜶
𝒔/ 𝒏
𝑨𝒄𝒄𝒆𝒑𝒕𝒂𝒃𝒍𝒆 𝒓𝒆𝒈𝒊𝒐𝒏

𝑹𝒆𝒋𝒆𝒄𝒕𝒊𝒐𝒏 𝒓𝒆𝒈𝒊𝒐𝒏

𝑹𝒆𝒋𝒆𝒄𝒕𝒊𝒐𝒏 𝒓𝒆𝒈𝒊𝒐𝒏

−𝒕𝜶/𝟐 t= 𝟎 𝒕𝜶/𝟐

ഥ from each term


Multiplying each term in the inequality by s/ 𝒏 and subtracting 𝒙
multiplying by ‘-1’
ഥ−𝝁
𝒙
ഥ − 𝒕𝜶/𝟐 𝒔/ 𝒏 <
𝑷 𝒙 ഥ + 𝒕𝜶/𝟐 𝒔/ 𝒏
<𝒙
𝒔/ 𝒏
Where 𝒙ഥ and 𝒔 is the mean and SD of the random sample of size ‘n’
from the population with known variance 𝒔𝟐
𝒕𝜶/𝟐 is the 𝒕 value leaving an area 𝜶/𝟐 to the right side. So,
Maximum error of estimate E with (𝟏 − 𝜶) probability is given by
𝒔
𝑬 = 𝒕𝜶/𝟐
𝒏
Where 𝜶, 𝑬, 𝝈 are known, then sample size ‘n’ is given by

𝟐
𝒔 𝒕𝜶/𝟐
𝒏=
𝑬
Example 1:
What is the size of the smallest sample required to estimate an
unknown proportion to within a maximum error of 0.06 with at
least 95% confidence.
Solution:
Given 𝑬 = 𝟎. 𝟎𝟔 and 𝒏 =?
We know that 𝟏 − 𝜶 𝟏𝟎𝟎 = 𝟗𝟓
𝟏 − 𝜶 = 𝟎. 𝟗𝟓
𝜶 = 𝟏 − 𝟎. 𝟗𝟓 = 𝟎. 𝟎𝟓
𝒁𝜶/𝟐 =
𝜶/𝟐 = 𝟎. 𝟎𝟐𝟓 𝟎. 𝟒𝟕𝟓
𝒛=𝟎
𝑷 𝒁𝜶/𝟐 = 𝟏. 𝟗𝟔 (from normal distribution table)
𝝈 𝒁𝜶/𝟐 𝟐 𝒁𝜶/𝟐 𝟐
Sample size, 𝒏 = = 𝒑𝒒
𝑬 𝑬
since 𝝈 = 𝒏𝒑𝒒 for max n=1. therefore 𝝈 = 𝒑𝒒
𝒁𝜶/𝟐 𝟐
Hence 𝒏 = 𝒑𝒒 = 𝟐𝟔𝟔. 𝟕𝟕 ≈ 𝟐𝟔𝟕
𝑬
Example 2:
If we assert with 95% that the maximum error is 0.05 and ‘p’
values is 0.2. Find the sample size
Solution:
Given that maximum error 𝑬 = 𝟎. 𝟎𝟓
We know that 𝟏 − 𝜶 𝟏𝟎𝟎 = 𝟗𝟓
𝟏 − 𝜶 = 𝟎. 𝟗𝟓
𝜶 = 𝟏 − 𝟎. 𝟗𝟓 = 𝟎. 𝟎𝟓 𝒁𝜶/𝟐 =
𝜶/𝟐 = 𝟎. 𝟎𝟐𝟓 𝟎. 𝟒𝟕𝟓
𝑷 𝒁𝜶/𝟐 = 𝟏. 𝟗𝟔 (from normal distribution table) 𝒛 = 𝟎
𝝈 𝒁𝜶/𝟐 𝟐 𝒁𝜶/𝟐 𝟐
Sample size, 𝒏 = = 𝒑𝒒
𝑬 𝑬
Given 𝒑 = 𝟎. 𝟐 ⇒ 𝒒 = 𝟎. 𝟖
𝒁𝜶/𝟐 𝟐
Hence 𝒏 = 𝒑𝒒 = 𝟐𝟒𝟓. 𝟖𝟔 ≈ 𝟐46
𝑬
Example 3:
It is desired to estimate the mean no.of hours of continuous use
until a certain computer will first require repairs. If it can be
assumed that 𝝈 value 48 hours. How large a sample be needed so
that one will be able to assert with 90% confidence that the
sample mean is of that atmost by 10 hours.
Solution:
Given 𝝈 = 𝟒𝟖 𝒉𝒐𝒖𝒓𝒔 𝒂𝒏𝒅 𝑬 = 𝟏𝟎
We know that 𝟏 − 𝜶 𝟏𝟎𝟎 = 𝟗0
𝟏 − 𝜶 = 𝟎. 𝟗0 𝒁𝜶/𝟐 =
𝜶 = 𝟏 − 𝟎. 𝟗𝟎 = 𝟎. 𝟎1 𝟎. 𝟒𝟓
𝜶/𝟐 = 𝟎. 𝟎𝟓 𝒛=𝟎
𝑷 𝒁𝜶/𝟐 = 𝟏.65 (from normal distribution table)
𝝈 𝒁𝜶/𝟐 𝟐
Sample size, 𝒏 = = 𝟔𝟐. 𝟕𝟐 ≈ 𝟔𝟑
𝑬
Example 4:
The mean and SD of a population are 11795 and 14054
respectively . What one can assert with 95% confidence about the
maximum error with n=50.
(or)
If the mean and SD of a population are 11795 and 14054. if n value
is 50. Find 95% confidence interval for the mean.
Solution:
Sample mean 𝒙 ഥ = 𝟏𝟏𝟕𝟗𝟓=Population mean 𝝁
SD =14054
We know that 𝟏 − 𝜶 𝟏𝟎𝟎 = 𝟗𝟓
𝟏 − 𝜶 = 𝟎. 𝟗𝟓
𝜶 = 𝟏 − 𝟎. 𝟗𝟓 = 𝟎. 𝟎𝟓 𝒁𝜶/𝟐 =
𝟎. 𝟒𝟕𝟓
𝜶/𝟐 = 𝟎. 𝟎𝟐𝟓
𝑷 𝒁𝜶/𝟐 = 𝟏. 𝟗𝟔 (from normal distribution table) 𝒛 = 𝟎
𝝈 𝝈
ഥ − 𝒁𝜶
Confidence interval= 𝒙 , ഥ + 𝒁𝜶
𝒙
𝟐 𝒏, 𝟐 𝒏,
𝟏𝟒𝟎𝟓𝟒 𝟏𝟒𝟎𝟓𝟒
𝟏𝟏𝟕𝟗𝟓 − 𝟏. 𝟗𝟔 , 𝟏𝟏𝟕𝟗𝟓 + 𝟏. 𝟗𝟔
𝟓𝟎 𝟓𝟎
𝟕𝟖𝟗𝟗. 𝟒, 𝟏𝟓𝟔𝟗𝟎. 𝟓𝟕

Practice problems
Example 5:
The efficiency expert of a computer company tested 40 engineers to
estimate the average time it takes to assemble a certain computer
component, getting a mean of 12.73 minutes and SD of 2.06 minutes
ഥ = 𝟏𝟐. 𝟕𝟑 is used as a point estimate of the actual value time
(a) If 𝒙
required to perform the task, determine the maximum error with
99% confidence.
(b) Construct 98% confidence interval for the true average time it
takes to do the job.
(c) With what confidence can we assert that the sample mean does
not differ from the true ,mean by more than 30 seconds
Example 6:
A sample of 10 cam shafts intended for use in gasoline engines has
an average eccentricity of 1.02 and the standard deviation of 0,044
inch. Assuming the data may be treated random sample from a
normal population, determine a 95% confidence interval for the
actual mean eccentricity of the cam shaft.

Example 7:
Determine a 95% confidence interval for the mean of a normal
distribution with variance 0.25, using a sample of n=100 values
with mean 212.3.

Example 7:
A random sample of 100 teachers in a large metropolitan area
revealed a mean weekly salary of Rs. 487 with SD Rs 48. With what
degree of confidence can we assert than the average weekly salary
of all teachers in the metropolitan area between 472 to 502.
Maximum likely hood estimator
A particular type of estimator, known as the maximum likelihood
estimator, is widely used in statistics.
෡ is defined to be that the value
The maximum likelihood estimate 𝜽
of 𝜽 maximizing 𝒇(𝒙𝟏 , 𝒙𝟐 , 𝒙𝟑 , 𝒙𝟒 …… 𝒙𝒏 |𝜽) where 𝒙𝟏 , 𝒙𝟐 , 𝒙𝟑 , 𝒙𝟒 ……
𝒙𝒏 are observed values. The function 𝒇(𝒙𝟏 , 𝒙𝟐 , 𝒙𝟑 , 𝒙𝟒 …… 𝒙𝒏 |𝜽)
often referred to as the likelihood function of 𝜽.
In determining the maximum value of 𝜽 it is often useful to use the
fact that 𝒇(𝒙𝟏 , 𝒙𝟐 , 𝒙𝟑 , 𝒙𝟒 …… 𝒙𝒏 |𝜽) and log ሼ𝒇(𝒙𝟏 , 𝒙𝟐 , 𝒙𝟑 , 𝒙𝟒 ……
𝒙𝒏 |𝜽) ሽ have their maximum at the same value of 𝜽.
෡ by maximizing
Hence, we may also obtain 𝜽
log 𝒇(𝒙𝟏 , 𝒙𝟐 , 𝒙𝟑 , 𝒙𝟒 …… 𝒙𝒏 |𝜽)
The maximum likelihood estimator of the unknown mean of a
Bernoulli distribution is given by
σ𝒏
𝒊=𝟏 𝑿𝒊 σ𝒏
𝒊=𝟏 𝑿𝒊
𝒅(𝒙𝟏 , 𝒙𝟐 , 𝒙𝟑 , 𝒙𝟒 …… 𝒙𝒏 )= ෝ=
also 𝒑
𝒏 𝒏
Since σ𝒏𝒊=𝟏 𝑿𝒊 is the number of successful trials, we see that the
maximum likelihood estimator of p is equal to the proportion of
the observed trials that result in successes.

The maximum likelihood estimator of the unknown mean of a


Poisson distribution is given by
σ𝒏
𝒊=𝟏 𝑿𝒊 σ𝒏
𝒊=𝟏 𝑿𝒊
𝒅(𝒙𝟏 , 𝒙𝟐 , 𝒙𝟑 , 𝒙𝟒 …… 𝒙𝒏 )= also 𝝀෠ =
𝒏 𝒏
Example 1:
The number of traffic accidents in Berkeley, California, in 10
randomly chosen non-rainy days in 1998 is as follows: 4, 0, 6, 5, 2, 1,
2, 0, 4, 3. Use these data to estimate the proportion of non-rainy
days that had 2 or fewer accidents that year.

Solution:
Use these data to estimate the proportion of non-rainy days that
had 2 or fewer accidents that year
𝟏𝟎
𝟏
ഥ=
𝑿 ෍ 𝑿𝒊 = 𝟐. 𝟕
𝟏𝟎
𝒊=𝟏
it follows that the maximum likelihood estimate of the Poisson
mean is 2.7. Since the long-run proportion of non-rainy days that
have 2 or fewer accidents is equal to 𝑷 𝑿 ≤ 𝟐
where X is the random number of accidents in a day, it follows
that the desired estimate is
𝑷 𝑿≤𝟐 =𝑷 𝑿=𝟎 +𝑷 𝑿=𝟏 +𝑷 𝑿=𝟐

𝟐. 𝟕 𝟐
= 𝒆−𝟐.𝟕 𝟏 + 𝟐. 𝟕 + = 𝟎. 𝟒𝟗𝟑𝟔
𝟐
Therefore, we estimate that a little less than half of the non-rainy
days had 2 or fewer accidents.

Example 2:
Two proofreaders were given the same manuscript to read. If
proofreader 1 found 𝒏𝟏 errors, and proofreader 2 found 𝒏𝟐 errors,
with 𝒏𝟏,𝟐 of these errors being found by both proofreaders,
estimate N, the total number of errors that are in the manuscript.
Before we can estimate N we need to make some assumptions
about the underlying probability model. So let us assume that the
results of the proofreaders are independent, and that each error in
the manuscript is independently found by proofreader 𝒊 with
probability 𝒑𝒊 , 𝒊 = 𝟏, 𝟐. To estimate N, we will start by deriving an
estimator of 𝒑𝟏 . To do so, note that each of the 𝒏𝟐 errors found by
reader 2 will, independently, be found by proofreader 1 with
probability pi. Because proofreader 1 found 𝒏𝟏,𝟐 of those 𝒏𝟐
errors, a reasonable estimate of 𝒑𝟏 is given by

𝒏𝟏,𝟐
ෝ=
𝒑
𝒏𝟐

You might also like