Chapter10 Sampling Two Stage Sampling
Chapter10 Sampling Two Stage Sampling
In cluster sampling, all the elements in the selected clusters are surveyed. Moreover, the efficiency in
cluster sampling depends on size of the cluster. As the size increases, the efficiency decreases. It
suggests that higher precision can be attained by distributing a given number of elements over a large
number of clusters and then by taking a small number of clusters and enumerating all elements within
them. This is achieved in subsampling.
In subsampling
- divide the population into clusters.
- Select a sample of clusters [first stage}
- From each of the selected cluster, select a sample of specified number of elements [second stage]
The clusters which form the units of sampling at the first stage are called the first stage units and the
units or group of units within clusters which form the unit of clusters are called the second stage units or
subunits.
The procedure is generalized to three or more stages and is then termed as multistage sampling.
Cluster sampling is a special case of two stage sampling in the sense that from a population of N
clusters of equal size m = M , a sample of n clusters are chosen.
If further M= m= 1, we get SRSWOR.
If n = N , we have the case of stratified sampling.
yij : Value of the characteristic under study for the j th second stage units of the i th first stage
=
unit; i 1,=
2,..., N ; j 1, 2,.., m.
M
1
Yi =
M
∑y
j =1
ij : mean per 2nd stage unit of i th 1st stage units in the population.
1 N M 1 N
=Y ∑∑ ij N=
=
MN =i 1 =j 1
y ∑ yi YMN : mean per second stage unit in the population
=i 1
1 m
yi = ∑
m j =1
yij : mean per second stage unit in the i th first stage unit in the sample.
1 n m 1 n
=y ∑∑ ij n=
=
mn=i 1 =j 1
y ∑ yi ymn : mean per second stage in the sample.
=i 1
Advantages:
The principle advantage of two stage sampling is that it is more flexible than the one stage sampling. It
reduces to one stage sampling when m = M but unless this is the best choice of m , we have the
opportunity of taking some smaller value that appears more efficient. As usual, this choice reduces to a
balance between statistical precision and cost. When units of the first stage agree very closely, then
consideration of precision suggests a small value of m . On the other hand, it is sometimes as cheap to
measure the whole of a unit as to a sample. For example, when the unit is a household and a single
respondent can give as accurate data as all the members of the household.
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 2
A pictorial scheme of two stage sampling scheme is as follows:
Second stage
sample
m units
Cluster Cluster Cluster n clusters
… … …
m units m units m units (large number
of elements
mn units from each
cluster)
Note: The expectations under two stage sampling scheme depend on the stages. For example, the
expectation at second stage unit will be dependent on first stage unit in the sense that second stage unit
will be in the sample provided it was selected in the first stage.
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 3
In case of two stage sampling,
{
E (θˆ) = E1 E2 E3 (θˆ) .
}
Consider
E2 (θˆ − θ )=
2
E2 (θˆ 2 ) − 2θ E2 (θˆ) + θ 2
{ }
.
= E2 (θˆ + V2 (θˆ) − 2θ E2 (θˆ) + θ 2
2
{ }
= E1 E2 (θˆ) − θ 2 + E1 V2 (θˆ)
2
= Var (θˆ) V1 E2 (θˆ) + E1 V2 (θˆ) .
{ } { } { }
Var (θˆ) = V1 E E3 (θˆ) + E1 V2 E3 (θˆ) + E1 E2 V3 (θˆ) .
2
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 4
Estimation of population mean:
Consider y = ymn as an estimator of the population mean Y .
Bias:
Consider
E ( y ) = E1 [ E2 ( ymn ) ]
=
E1 E2 ( yim i ) (as 2nd stage is dependent on 1st stage)
=
E1 E2 ( yim i ) (as yi is unbiased for Yi due to SRSWOR)
1 n
= E1 ∑ Yi
n i =1
1 N
= ∑ Yi
N i =1
=Y .
Variance
) E1 V2 ( y i ) + V1 [ E2 ( y / i ) ]
Var ( y =
1 n 1 n
E1 V2 ∑ yi i + V1 E2 ∑ yi / i
= n i 1= n i 1
1 n 1 n
E1 2 ∑ V ( yi i ) + V1 ∑ E2 ( yi / i )
= n i 1= n i 1
1 n
1 1 1 n
= E1 2 ∑ − Si2 + V1 ∑ Yi
= m M
n i 1= n i 1
1 n 1 1
2 ∑
= − E1 ( Si2 ) + V1 ( yc )
n i =1 m M
(where yc is based on cluster means as in cluster sampling)
1 1
1 N −n 2
= n − S w2 + Sb
m M
n2 Nn
1 1 1 2 1 1 2
= − S w + − Sb
nm M n N
2
1 N 2
(Yij − Yi )
N M
1
=
where S w2 = ∑ i N (M − 1)=∑∑
N =i 1
S
i 1 =j 1
1 N
=S ∑
N − 1 i =1
b(Yi − Y ) 2
2
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 5
Estimate of variance
An unbiased estimator of variance of y can be obtained by replacing Sb2 and S w2 by their unbiased
Consider an estimator of
N
1
S w2 =
N
∑S
i =1
i
2
∑ ( yij − Yi )
1 M
=
where Si2
M − 1 j =1
1 n 2
as ∑ si
sw2 =
n i =1
1 m
=
where si2 ∑ ( yij − yi )2 .
m − 1 j =1
So
E ( sw2 ) = E1 E2 ( sw2 i )
1 n
= E1 E2 ∑ si2 i
n i =1
n
1
= E1 ∑ E2 ( si2 i )
n i =1
1 n 2
= E1 ∑ Si
n i =1
(as SRSWOR is used)
1 n
= ∑
n i =1
E1 ( Si2 )
1 N 1 N 2
= ∑ N ∑
N i 1=
= i 1
Si
N
1
=
N
∑S
i =1
i
2
=S 2
w
Consider
1 n
=sb2 ∑
n − 1 i =1
( yi − y ) 2
as an estimator of
1 N
=Sb2 ∑
N − 1 i =1
(Yi − Y ) 2 .
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 6
So
1 n
=E ( sb2 ) E ∑ ( yi − y ) 2
n − 1 i =1
n
(n − 1) E ( sb2 ) = E ∑ yi2 − ny 2
i =1
n
= E ∑ yi2 − nE ( y 2 )
i =1
n
= E1 E2 ∑ yi2 − n Var ( y ) + { E ( y )}
2
i =1
n 1 1 1 1 1
= E1 ∑ E2 ( yi2 ) i ) − n − Sb2 + − S w2 + Y 2
i =1 n N m M n
n
{
i =1
}
2
1 1 1 1 1
= E1 ∑ Var ( yi ) + ( E ( yi ) − n − Sb2 + − S w2 + Y 2
n N m M n
n 1 1 2 2 1 1 2 1 1 1 2 2
= E1 ∑ − Si + Yi − n − Sb + − Sw + Y
i =1 m M n N m M n
1 n 1 1 2 2 1 1 2 1 1 1 2 2
= nE1 ∑ − Si + Yi − n − Sb + − Sw + Y
n i =1 m M n N m M n
1 1 1 N 1 N 1 1 1 1 1
= n − ∑ Si2 + ∑ Yi 2 − n − Sb2 + − S w2 + Y 2
m =
M N i 1= N i1 n N m M n
1 1 1 N
1 1 1 1 1
= n − S w2 + ∑ Yi 2 − n − Sb2 + − S w2 + Y 2
m M N i =1 n N m M n
1 1 2 n N 2 1 1 2
=(n − 1) − S w + ∑ Yi − nY − n − Sb
2
m M N i =1 n N
1 1 2 nN 2 2 1 1 2
=(n − 1) − S w + ∑ Yi − NY − n − Sb
m M N i =1 n N
1 1 n 1 1
= (n − 1) − S w2 + ( N − 1) Sb2 − n − Sb2
m M N n N
1 1
= (n − 1) − S w2 + (n − 1) Sb2 .
m M
1 1 2
⇒ E ( sb2 ) = − S w + Sb
2
m M
1 1
or E sb2 − − sw2 = Sb2 .
m M
Thus
( y )= 1 1 − 1 Sˆ 2 + 1 − 1 Sˆ 2
Var ω b
nm M n N
1 1 1 2 1 1 2 1 1 2
= − sw + − sb − − sw
nm M n N m M
11 1 2 1 1 2
= − sw + − sb .
N m M n N
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 7
Allocation of sample to the two stages: Equal first stage units:
The variance of sample mean in the case of two stage sampling is
( y )= 1 1 − 1 S 2 + 1 − 1 S 2 .
Var w b
nm M n N
It depends on Sb2 , S w2 , n and m. So the cost of survey of units in the two stage sample depends on
n and m.
S2
This variance is monotonic decreasing function of n if Sb2 − w > 0. The variance is minimum
M
C0
increasing function of n , It reaches minimum when n assumes the minimum value, i.e., nˆ =
kM
(i.e., no subsampling).
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 8
(II) When cost function is =
C k1n + k2 mn
We have
S2 S2 S2 k S2
C0 Var ( y ) + b = k1 Sb2 − w + k2 S w2 + mk2 Sb2 − w + 1 w .
N M M m
S2
When Sb2 − w > 0, then
M
2 2
S2 S2 S2 k S2
C0 Var ( y ) + b= k1 Sb2 − w + k2 S w2 + mk2 Sb2 − w − 1 w
N M M m
which is minimum when the second term of right hand side is zero. So we obtain
k1 S w2
mˆ = .
k2 2 S w2
Sb −
M
C0
nˆ = .
k1 + k2 mˆ
S2
When Sb2 − w ≤ 0 then
M
S2 S2 S2 k S2
C0 Var ( y ) + b = k1 Sb2 − w + k2 S w2 + mk2 Sb2 − w + 1 w
N M M m
If N is large, then S w2 ≈ S 2 (1 − ρ )
S w2
S w2 − ≈ ρS2
M
k1 1
mˆ ≈ − 1 .
k2 ρ
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 9
Case 2: When variance is fixed
Now we find the sample sizes when variance is fixed, say as V0 .
1 1 1 2 1 1 2
V0= − S w + − Sb
nm M n N
1 1
Sb2 + − S w2
⇒n=
m M
Sb2
V0 +
N
So
2 S w2
Sb − kS w2
= = km
C kmn M + .
2 2
V + Sb V + Sb
0 0
N N
2 S w2
If Sb − > 0, C attains minimum when m assumes the smallest integral value, i.e., 1.
M
S2
If Sb2 − w < 0 , C attains minimum when mˆ = M .
M
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 10
The intraclass correlation coefficient is
N − 1 2 Sw
2
Sb −
M M ( N − 1) Sb − NS w
2 2
=ρ N
= ;−
1
≤ ρ ≤1 (1)
NM − 1 2 ( MN − 1) S 2
M −1
S
NM
and using the identity
N M N M N M
∑∑ ( yij − Y=
=i 1 =j 1
)2 ∑∑ ( yij − Yi )2 + ∑∑ (Yi − Y )2
=i 1 =j 1 =i 1 =j 1
( NM − 1) S =( N − 1) MS + N ( M − 1) S w2
2 2
b (2)
1 N M 1 M
=
where Y = ∑∑ ij i M =∑
MN =i 1 =j 1
y , Y
j 1
yij .
Now we need to find Sb2 and S w2 from (1) and (2) in terms of S 2 . From (1), we have
MN − 1 N −1
S w2 =
− MS ρ +
2 2
MSb . (3)
N N
N N
= ( N − 1) MSb + ( M − 1)( N − 1) Sb − ρ M ( M − 1)( MN − 1) S
2 2 2
N ( M − 1) S w2= ( NM − 1) S 2 − ( N − 1) MSb2
( MN − 1) S 2
= ( NM − 1) S 2 − ( N − 1) M 2 [1 + ( M − 1) ρ ]
M ( N − 1)
M − 1 − ( M − 1) ρ
= ( NM − 1) S 2
M
= ( NM − 1) S 2 ( M − 1)(1 − ρ )
MN − 1 2
⇒=S w2 S (1 − ρ ).
MN
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 11
Substituting Sb2 and S w2 in Var ( yTS )
MN − 1 S m(n − 1) N −n m M − m
2
=
V ( yTS ) 1− +ρ ( M − 1) − .
MN mn M ( N − 1) N −1 M M
m
When subsampling rate is small, MN − 1 ≈ MN and M − 1 ≈ M , then
M
S2
V ( ySRS ) =
mn
S2 N −n
V ( yTS ) =1 + ρ m − 1 .
mn N −1
The relative efficiency of the two stage in relation to one stage sampling of SRSWOR is
Var ( yTS ) N −n
RE = 1+ ρ
= m − 1 .
Var ( ySRS ) N −1
N −n N −n
If N − 1 ≈ N and finite population correction is ignorable, then ≈ ≈ 1, then
N −1 N
1 + ρ (m − 1).
RE =
1M 2 1 2
− 1 Sb − S w
n m M
which is approximately
1M 2 2 S w2
− 1 ρ S for large N and Sb − > 0.
n m M
MN − 1 S 2
=
where Sb2 [1 + ρ ( M − 1)]
M ( N − 1) M
MN − 1 2
=S w2 S (1 − ρ )
MN
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 12
So smaller the m / M , larger the reduction in the variance of two stage sample over a cluster sample.
S2
When Sb2 − w < 0 then the subsampling will lead to loss in precision.
M
mi : number of second stage units to be selected from i th first stage unit, if it is in the sample.
n
m0 = ∑ mi : total number of second stage units in the sample.
i =1
mi
1
yi ( mi ) =
mi
∑yj =1
ij
Mi
1
Yi =
Mi
∑y j =1
ij
1 N
=Y = ∑ yi
N i =1
YN
N Mi N
∑∑ yij ∑M Y
1 i i N
=Y =i 1 =j 1
=N
=
i =1
∑u Y i i
∑ Mi
MN N i =1
i =1
Mi
ui =
M
1 N
M = ∑ Mi
N i =1
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 13
The pictorial scheme of two stage sampling with unequal first stage units case is as follows:
N clusters
Second stage
Cluster Cluster Cluster sample
m1 units m2 units … … … mn units n clusters
(small)
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 14
Now we consider different estimators for the estimation of population mean.
1. Estimator based on the first stage unit means in the sample:
1 n
=ˆ
Y y=
S2 ∑ yi ( mi )
n i =1
Bias:
1 n
E ( yS 2 ) = E ∑ yi ( mi )
n i =1
1 n
= E1 ∑ E2 ( yi ( mi ) )
n i =1
1 n
= E1 ∑ Yi [Since a sample of size mi is selected out of M i units by SRSWOR]
n i =1
1 N
= ∑ Yi
N i =1
=YN
≠ Y.
Bias=
( yS 2 ) E ( yS 2 ) − Y
1 N 1 N
=
=
∑ i NM ∑
Y
N i 1=
−
i 1
M iYi
1 N 1 N N
=
− ∑ M iYi − ∑ Yi ∑ M i
NM
= i 1 N i 1=
= i 1
N
1
= ∑ (M i − M )(Yi − YN ).
NM i =1
This bias can be estimated by
(y ) = N −1 n
Bias S2 − ∑ (M i − m)( yi ( mi ) − yS 2 )
NM (n − 1) i =1
which can be seen as follows:
N −1 1 n
( y ) =
E Bias − E1 ∑ E2 {( M i − m)( yi ( mi ) − yS 2 ) / n}
S 2 NM n − 1 i =1
N −1 1 n
=
− E ∑
NM n − 1 i =1
( M i − m)(Yi − yn )
N
1
=
−
NM
∑ (M
i =1
i − M )(Yi − YN )
= YN − Y
1 n
where yn = ∑ Yi .
n i =1
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 15
An unbiased estimator of the population mean Y is thus obtained as
N −1 1 n
yS 2 + ∑ (M i − m)( yi ( mi ) − yS 2 ) .
NM n − 1 i =1
Note that the bias arises due to the inequality of sizes of the first stage units and probability of selection
of second stage units varies from one first stage to another.
Variance:
Var ( yS=
2) Var E ( yS 2 n) + E Var ( yS 2 n)
1 n 1 n
= Var ∑ yi + E 2 ∑ Var ( yi ( mi ) i )
= n i 1= n i 1
1 1 2 1 n
1 1 2
=
− Sb + E 2
n N
∑ m − Si
n i =1 i Mi
1 1 2 1 N 1 1 2
=
− Sb +
n N
∑ −
Nn i =1 mi M i
Si
( )
2
1 N
=
where S 2
∑ Yi − YN
N − 1 i =1
b
2
1 Mi
=S ∑ ( yij − Yi ) .
2
M i − 1 j =1
i
= ( yS 2 ) Var ( yS 2 ) + [ Bias ( yS 2 ) ] .
2
MSE
Estimation of variance:
Consider mean square between cluster means in the sample
1 n
∑ ( )
2
=sb2 yi ( mi ) − yS 2 .
n − 1 i =1
It can be shown that
1 N
1 1 2
E ( sb2 ) =
Sb2 + ∑ m − Si .
N i =1 i Mi
1 mi
=
Also si2 ∑
mi − 1 j =1
( yij − yi ( mi ) ) 2
1 Mi
E( s =
2
i) S= 2
∑
M i − 1 j =1
i ( yij − Yi ) 2
1 n 1 1 2 1 N 1 1 2
So E ∑ − si = ∑ − Si .
n i 1 i
=m M i N i 1 i
m M i
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 16
Thus
1 n 1 1 2
Sb2 + E ∑ −
E ( sb2 ) = si
n i =1 i
m M i
and an unbiased estimator of Sb2 is
1 n 1 1 2
Sb =
ˆ 2
sb − ∑ −
2
si .
n i =1 mi M i
So an estimator of the variance can be obtained by replacing Sb2 and Si2 by their unbiased estimators as
(y ) = 1 1 ˆ2 1 N 1 1 ˆ2
Var S2 − Sb +
n N
∑ −
Nn i =1 mi M i
Si .
=Y.
Variance:
Var ( yS* 2 ) Var E ( yS* 2 n) + E Var ( yS* 2 n)
=
1 n 1 n
= Var ∑ uiYi + E 2 ∑ ui2Var ( yi ( mi ) i )
= n i 1= n i 1
1 1 *2 1 N
1 1 2
=−
n N
Sb + ∑u 2
i − Si
nN i =1 mi M i
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 17
1 Mi
=
where Si2 ∑
M i − 1 j =1
( yij − Yi ) 2
1 N
=Sb*2 ∑
N − 1 j =1
(uiYi − Y ) 2 .
∑M y i i ( mi )
Yˆ y=
= **
S2
i =1
n
∑M
i =1
i
∑u y i i ( mi )
= i =1
n
∑ui =1
i
yS* 2
=
un
Mi 1 n
=
where ui =
M
, un ∑ ui .
n i =1
This estimator can be seen as if arising by the ratio method of estimation as follows:
Let yi* = ui yi ( mi )
Mi
=xi* = , i 1, 2,..., N
M
be the values of study variable and auxiliary variable in reference to the ratio method of estimation. Then
1 n *
=y* =∑
n i =1
yi yS* 2
1 n *
=x* = ∑ xi un
n i =1
1 N *
=X* = ∑ X i 1.
N i =1
So the bias and mean squared error of yS**2 can be obtained directly from the results of ratio estimator.
Recall that in ratio method of estimation, the bias and MSE of the ratio estimator upto second order of
approximation is
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 18
N −n
Bias ( yˆ R ) ≈ Y (C x2 − 2 ρ C xC y )
Nn
Var ( x ) Cov( x , y )
= Y −
X
2
XY
MSE (YˆR ) ≈ Var ( y ) + R 2Var ( x ) − 2 RCov( x , y )
Y
where R = .
X
Bias:
The bias of yS**2 up to second order of approximation is
1 n 1 n 1 n 1 n
Cov( xS*2 , yS* 2 ) Cov E ∑ ui xi ( mi ) , ∑ ui yi ( mi ) + E Cov ∑ ui xi ( mi ) , ∑ ui yi ( mi )
= n i 1= ni1 = n i 1= ni1
1 n 1 n 1 n
Cov ∑ ui E ( xi ( mi ) ), ∑ ui E ( yi ( mi ) ) + E 2 ∑ ui2Cov( xi ( mi ) , yi ( mi ) ) i
= n i 1= ni1 = n i 1
1 n 1 n 1 n 1 1
= Cov ∑ ui X i , ∑ uiYi + E 2 ∑ ui2 − Sixy
= n i 1= ni1 = n i 1 mi M i
1 1 * 1 N
1 1
=
− Sbxy +
n N
∑u 2
i − Sixy
nN i =1 mi M i
where
1 N
=*
Sbxy ∑ (ui X i − X )(uiYi − Y )
N − 1 i =1
1 Mi
=
Sixy ∑ ( xij − X i )( yij − Yi ).
M i − 1 j =1
Similarly, Var ( xS*2 ) can be obtained by replacing x in place of y in Cov( xS*2 , yS* 2 ) as
1 1 *2 1 N
1 1 2
Var ( xS*2 ) =−
n N
Sbx + ∑u 2
i − Six
nN i =1 mi M i
1 N
=
where Sbx*2 ∑ (ui X i − X )2
N − 1 i =1
1 Mi
=Six*2 ∑ ( xij − X i )2 .
M i − 1 i =1
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 19
Substituting Cov( xS*2 , yS* 2 ) and Var ( xS*2 ) in Bias ( yS**2 ), we obtain the approximate bias as
1 1 Sbx*2 Sbxy
*
1 N
1 1 Six2 Sixy
Bias ( y ) ≈ Y − 2 −
**
S2 +
XY nN
∑ u 2
i − 2 − .
n N X i =1 mi M i X XY
1 1 1 1 2
MSE ( yS**2 ) ≈ − ( Sby*2 − 2 R* Sbxy + R*2 Sbx*2 ) + ( Siy − 2 R Sixy + R Six ) .
N
1
n N
*
∑ u 2
i −
* *2 2
nN i =1 mi M i
Also
1 1 1 N 2 1 N 2 1 1 2
∑ ui (Yi − R* X i ) + ∑ ( Siy − 2 R Sixy + R Six ) .
2
MSE ( yS**2 ) ≈ − ui −
* *2 2
= n N N −1 i 1= nN i 1 mi M i
Estimate of variance
Consider
( ui yi ( mi ) − yS* 2 )( ui xi ( mi ) − xS*2 )
1 n
s=
*
bxy ∑
n − 1 i =1
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 20
It can be shown that
1 1
E ( sbxy )= 1 N
* *
Sbxy + ∑ ui2 − Sixy
N i =1 mi M i
E ( sixy ) = Sixy .
So
1 n 1 1 1 N 2 1 1
E ∑ ui2 − sixy=
∑ ui − Sixy .
n i 1= mi M i N i 1 mi M i
Thus
1 n 1 1
Sˆbxy
*
= *
sbxy − ∑ ui2 − sixy
n i =1 mi M i
1 n 1 1 2
Sˆbx*2 = *2
sbx − ∑ ui2 − six
n i =1 mi M i
1 n 1 1 2
Sˆby*2 =*2
sby − ∑ ui2 − siy .
n i =1 mi M i
Also
1 n 1 1 2 1 N 2 1 1 2
E ∑ ui2 − six =
∑ ui − Six
n i 1 = mi M i N i 1 mi M i
1 n 1 1 2 1 N 2 1 1 2
E ∑ ui2 − siy =
∑ ui − Siy .
n i 1 = mi M i N i 1 mi M i
A consistent estimator of MSE of yS**2 can be obtained by substituting the unbiased estimators of
( y ** ) ≈ 1 − 1 ( s*2 − 2r * s* + r *2 s*2 )
MSE S2 by bxy bx
n N
1 n 2 1 1 2
+ ∑ ui − ( siy − 2r sixy + r six )
* *2 2
nN i =1 mi M i
1 1 1 n
∑ ( yi ( mi ) − r * xi ( mi ) )
2
≈ −
n N n − 1 i =1
1 n 2 1 1 2 yS* 2
+ ∑ ui − ( siy − 2r sixy + r six ) where r * =
* *2 2
.
nN i =1 mi M i xS*2
Sampling Theory| Chapter 10 | Two Stage Sampling | Shalabh, IIT Kanpur Page 21