Sample Size
Sample Size
Introduction
Larger sample sizes generally lead to increased precision when estimating unknown parameters. For example, if we wish to know the proportion of a certain species of fish that is infected with a pathogen, we would generally have a more accurate estimate of this proportion if we sampled and examined 200, rather than 100 fish. Several fundamental facts of mathematical statistics describe this phenomenon, including the law of large numbers and the central limit theorem. In some situations, the increase in accuracy for larger sample sizes is minimal, or even non-existent. This can result from the presence of systematic errors or strong dependence in the data, or if the data follow a heavy-tailed distribution. Sample sizes are judged based on the quality of the resulting estimates. For example, if a proportion is being estimated, one may wish to have the 95% confidence interval be less than 0.06 units wide. Alternatively, sample size may be assessed based on the power of a hypothesis test. For example, if we are comparing the support for a certain political candidate among women with the support for that candidate among men, we may wish to have 80% power to detect a difference in the support levels of 0.04 units.
people out of the n sampled people who are at least 65 years old). When the observations are independent, this estimator has a (scaled) binomial distribution (and is also the sample mean of data from a Bernoulli distribution). The maximum variance of this distribution is 0.25/n, which occurs when the true parameter is p = 0.5. In practice, since p is unknown, the maximum variance is often used to for sample size assessments. For sufficiently large n, the distribution of
[1]
mean and variance. Using this approximation, it can be shown that around 95% of this distribution's probability lies within 2 standard deviations of the mean. Because of this, an interval of the form
will form a 95% confidence interval for the true proportion. If this interval needs to be no more than W units wide, the equation
can be solved for n, yielding[2] [3] n=4/W2=1/B2 where B is the error bound on the estimate, i.e., the estimate is usually given as within B. So, for B = 10% one requires n = 100, for B = 5% one needs n = 400, for B = 3% the requirement approximates to n = 1000, while for B = 1% a sample size of n = 10000 is required. These numbers are quoted often in news reports of opinion polls and other sample surveys.
Estimation of means
A proportion is a special case of a mean. When estimating the population mean using an independent and identically distributed (iid) sample of size n, where each data value has variance 2, the standard error of the sample mean is:
This expression describes quantitatively how the estimate becomes more precise as the sample size increases. Using the central limit theorem to justify approximating the sample mean with a normal distribution yields an approximate 95% confidence interval of the form
If we wish to have a confidence interval that is W units in width, we would solve for n, yielding the sample size n=162/W2. For example, if we are interested in estimating the amount by which a drug lowers a subject's blood pressure with a confidence interval that is six units wide, and we know that the standard deviation of blood pressure in the population is 15, then the required sample size is 100.
By tables
[4] Power 0.25 0.50 0.60 0.70 0.80 0.90 0.95 0.99 Cohen's d 0.2 84 193 246 310 393 526 651 920 0.5 14 32 40 50 64 85 105 148 0.8 6 13 16 20 26 34 42 58
Sample size determination The table shown at right can be used to in a two-sample t-test to estimate the sample sizes of an experimental group and a control group that are of equal size, that is, the total number of individuals in the trial is twice that of the number given, and the desired significance level is 0.05.[4] The parameters used are: The desired statistical power of the trial, shown in column to the left. Cohen's d, which is the expected difference between the means of the target values between the experimental group and the control group, divided by the expected standard deviation.
where: N is the total number of individuals or units in the study (minus 1) B is the blocking component, representing environmental effects allowed for in the design (minus 1) T is the treatment component, corresponding to the number of treatment groups (including control group) being used, or the number of questions being asked (minus 1) E is the degrees of freedom of the error component, and should be somewhere between 10 and 20. For example, if a study using laboratory animals is planned with four treatment groups (T=3), with eight animals per group, making 32 animals total (N=31), without any further stratification (B=0), then E would equal 28, which is above the cutoff of 20, indicating that sample size may be a bit too large, and six animals per group might be more appropriate.[6]
and an alternative hypothesis: for some 'smallest significant difference' * >0. This is the smallest value for which we care about observing a difference. Now, if we wish to (1) reject H0 with a probability of at least 1- when Ha is true (i.e. a power of 1-), and (2) reject H0 with probability when H0 is true, then we need the following: If z is the upper percentage point of the standard normal distribution, then and so 'Reject H0 if our sample average ( ) is more than '
is a decision rule which satisfies (2). (Note, this is a 1-tailed test) Now we wish for this to happen with a probability at least 1- when Ha is true. In this case, our sample average will come from a Normal distribution with mean *. Therefore we require
where
with
[8]
The weights, W(h), frequently, but not always, represent the proportions of the population elements in the strata, and W(h)=N(h)/N. For a fixed sample size, that is n=Sum{n(h)},
[9]
which can be made a minimum if the sampling rate within each stratum is made proportional to the standard deviation within each stratum: . An "optimum allocation" is reached when the sampling rates within the strata are made directly proportional to the standard deviations within the strata and inversely proportional to the square roots of the costs per element within the strata:
[10]
Notes
[1] NIST/SEMATECH, "7.2.4.2. Sample sizes required" (http:/ / www. itl. nist. gov/ div898/ handbook/ prc/ section2/ prc242. htm), e-Handbook of Statistical Methods. [2] "Large Sample Estimation of a Population Proportion" (http:/ / www. utdallas. edu/ ~ammann/ stat3355/ node25. html) [3] "Confidence Interval for a Proportion" (http:/ / nebula. deanza. fhda. edu/ ~bloom/ Math10/ M10ConfIntNotes. pdf) [4] Chapter 13 (http:/ / davidakenny. net/ statbook/ chapter_13. pdf), page 215, in: Kenny, David A. (1987). Statistics for the social and behavioral sciences. Boston: Little, Brown. ISBN0-316-48915-8. [5] Kirkwood, James; Robert Hubrecht (2010). The UFAW Handbook on the Care and Management of Laboratory and Other Research Animals. Wiley-Blackwell. pp.29. ISBN1-4051-7523-0. online Page 29 (http:/ / books. google. se/ books?id=Wjr9u1AAht4C& pg=PA29) [6] Isogenic.info > Resource equation (http:/ / www. isogenic. info/ html/ resource_equation. html) by Michael FW Festing. Updated Sept. 2006 [7] Kish (1965, Section 3.1) [8] Kish (1965), p.78. [9] Kish (1965), p.81. [10] Kish (1965), p.93. [11] Kish (1965), p.94.
References
Bartlett, J. E., II, Kotrlik, J. W., & Higgins, C. (2001). "Organizational research: Determining appropriate sample size for survey research" (http://www.osra.org/itlpj/bartlettkotrlikhiggins.pdf), Information Technology, Learning, and Performance Journal, 19(1) 43-50. Kish, L. (1965), Survey Sampling, Wiley. ISBN 047148900x
Further reading
NIST: Selecting Sample Sizes (http://www.itl.nist.gov/div898/handbook/ppc/section3/ppc333.htm) Raven Analytics: Sample Size Calculations (http://ravenanalytics.com/Articles/Sample_Size_Calculations. htm) ASTM E122-07: Standard Practice for Calculating Sample Size to Estimate, With Specified Precision, the Average for a Characteristic of a Lot or Process
License
Creative Commons Attribution-Share Alike 3.0 Unported http:/ / creativecommons. org/ licenses/ by-sa/ 3. 0/