Sampling and Hypothesis Testing
Sampling and Hypothesis Testing
Sampling
1. Population (N):
○ Definition: A complete set of elements or units relevant to the
research problem.
○ Examples: Individuals, households, organizations, villages, states, or
nations.
○ Enumeration: In some cases, the population can be fully listed
(enumerated), such as a voter list in a constituency.
2. Sample (n):
○ Definition: A finite subset of the population selected for study.
○ Characteristics: Must possess properties representative of the
entire population to ensure valid generalizations.
○ Elements: Can include individuals, households, organizations, etc.
3. Sampling Error:
○ Definition: The discrepancy between sample statistics and actual
population parameters due to studying only a part of the
population.
○ Causes: Natural variability, incorrect sampling procedures,
inadequate sample size, and non-representative samples.
○ Measurement: Assessed using the standard error, which quantifies
the variation of sample estimates around the population parameter.
○ Relationship with Sample Size: Inversely related; larger samples
typically reduce sampling error.
4. Statistic:
○ Definition: A summary measure derived from sample data.
○ Examples: Mean, median, mode, variance.
○ Purpose: Used to estimate population parameters.
5. Parameter:
○ Definition: A summary measure that describes a characteristic of
the entire population.
○ Examples: Population mean, population median, population mode.
○ Goal: Researchers aim to estimate these parameters using sample
statistics.
6. Estimate:
○ Definition: A value derived from sample data intended to
approximate a population parameter.
○ Unbiased Estimator: When the expected value of the estimator
equals the true population parameter.
○ Biased Estimator: When there is a systematic deviation between
the estimator and the parameter.
○ Bias: The difference between the expected estimator value and the
true population parameter.
7. Sampling Frame:
○ Definition: A list or representation of all elements in the population
from which the sample is drawn.
○ Development: Researchers may need to create a sampling frame
using various sources like censuses, records, or directories.
○ Example: A voter list for selecting participants in a political survey.
The universe of study refers to the totality of all possible elements or units that
fit the research criteria within a specified context. Properly defining this
universe is crucial for the validity and reliability of research findings.
Sampling techniques are broadly categorized into two main types: Probability
Sampling and Non-Probability Sampling.
A. Probability Sampling
Advantages:
● Representativeness: Enhances the likelihood that the sample mirrors the
population.
● Statistical Validity: Facilitates the use of statistical techniques to
estimate population parameters.
● Reduced Bias: Minimizes the influence of researcher bias in sample
selection.
Disadvantages:
B. Non-Probability Sampling
Advantages:
Disadvantages:
● Higher Risk of Bias: Increased potential for selection bias, which can
compromise the representativeness of the sample.
● Limited Generalizability: Findings may not be applicable to the entire
population.
● Subjectivity: Relies heavily on the researcher’s judgment, which can
introduce personal biases.
Methods of Implementation:
● Lottery Method:
○ Procedure:
1. Assign a unique number to each unit in the population.
2. Write these numbers on small pieces of paper (chits).
3. Place all chits in a container.
4. Draw the desired number of chits randomly to form the
sample.
● Tippet’s Table:
○ Procedure:
1. Assign unique numbers to each unit in the population.
2. Use a pre-generated table of random numbers (such as
Tippet’s table) to select sample units.
3. Match the numbers from the table to the assigned population
numbers.
Advantages:
Disadvantages:
2. Interval Sampling
Procedure:
1. List Preparation: Ensure all population units are listed in a specific order
(e.g., alphabetical, numerical).
2. Random Start: Use a random number generator or other randomization
method to select a starting point within the first interval.
3. Regular Interval Selection: Choose every 'n-th' unit from the starting
point to form the sample.
Advantages:
Disadvantages:
3. Stratified Sampling
Procedure:
Types:
● Stratified Random Sampling: Samples are randomly selected from each
stratum, maintaining the proportion of each stratum in the sample.
● Stratified Proportional Sampling: Samples are selected in proportion to
the size of each stratum in the population, ensuring proportional
representation.
Advantages:
Disadvantages:
4. Purposive Sampling
Characteristics:
● Non-Random Selection: Units are chosen intentionally rather than by
chance.
● Focus on Specific Criteria: Selection is based on predefined criteria
relevant to the research objectives, ensuring that the sample possesses
particular characteristics of interest.
Advantages:
Disadvantages:
● Bias Risk: High potential for researcher bias in sample selection, which
can compromise the objectivity of the study.
● Limited Generalizability: Findings may not be applicable to the broader
population due to the non-random selection of units.
● Questionable Representativeness: The sample may not accurately reflect
the diversity of the population, limiting the study’s overall validity.
5. Convenience Sampling
Characteristics:
● Ease of Access: Prioritizes accessibility and proximity over random
selection, focusing on units that are most convenient to reach.
● Non-Random: Selection is based on convenience rather than statistical
criteria, leading to potential biases.
Advantages:
Disadvantages:
● High Bias Potential: Sample may not represent the population accurately,
leading to skewed results.
● Limited Generalizability: Findings are often not applicable beyond the
sampled units, reducing the study’s external validity.
● Dependence on Researcher’s Judgment: Selection may be influenced by
personal biases, further compromising the objectivity of the study.
6. Cluster Sampling
Procedure:
1. Cluster Formation: Divide the population into distinct clusters (e.g.,
geographical regions, departments).
2. Cluster Selection: Randomly select a subset of these clusters.
3. Unit Selection: Study all units within the selected clusters or randomly
sample within them.
Advantages:
Disadvantages:
Advantages:
Disadvantages:
8. Quota Sampling
Procedure:
Advantages:
Disadvantages:
9. Multi-Stage Sampling
Definition: Multi-Stage Sampling is a complex form of cluster sampling where
the sampling process occurs in multiple stages, often combining different
sampling techniques at each stage. This method is particularly useful for
large-scale studies involving vast and dispersed populations.
Procedure:
Advantages:
Disadvantages:
Procedure:
Advantages:
Disadvantages:
Characteristics:
Advantages:
Disadvantages:
Formation of Hypotheses
Types of Hypotheses
A. Test Statistic
● Definition: A numerical value calculated from sample data, used to decide
whether to reject the null hypothesis.
● Function: Converts the observed data into a standardized form under the
assumption that the null hypothesis is true.
● Types: Depending on the research design and data type, common test
statistics include t-values, z-values, chi-square values, and F-statistics.
C. p-Value
D. Critical Value
● Definition: The threshold value that separates the rejection region from
the non-rejection region in the test statistic distribution.
● Determination: Based on the significance level and the distribution of the
test statistic.
● Usage: If the test statistic exceeds the critical value, the null hypothesis is
rejected.
E. Rejection Region
● Definition: The range of values for the test statistic that leads to the
rejection of the null hypothesis.
● Characteristics: Determined by the significance level and the nature of
the hypothesis (one-tailed or two-tailed).
● Implications: If the test statistic falls within the rejection region, the null
hypothesis is rejected.
● Definition: The range of values for the test statistic that leads to the
acceptance (or failure to reject) of the null hypothesis.
● Characteristics: Complementary to the rejection region.
● Implications: If the test statistic falls within this region, there is
insufficient evidence to reject the null hypothesis.
● Decision Rule:
○ If p ≤ α: Reject the null hypothesis (support the alternative
hypothesis).
○ If p > α: Fail to reject the null hypothesis (insufficient evidence to
support the alternative hypothesis).
● Interpretation:
○ Statistically Significant: If p ≤ α, indicating that the observed effect
is unlikely to have occurred by chance.
○ Not Statistically Significant: If p > α, indicating insufficient evidence
to support the observed effect.
● Components of Reporting:
○ Descriptive Statistics: Provide means, standard deviations, and
other relevant statistics.
○ Test Results: Present the test statistic, degrees of freedom (if
applicable), p-value, and critical values.
○ Interpretation: Explain the findings in the context of the research
question, discussing whether the hypotheses were supported.
● Example Report: "An independent samples t-test revealed that students
taught using the new interactive method (M = 85, SD = 5) scored
significantly higher on mathematics tests than those taught using
traditional methods (M = 78, SD = 6), t(98) = 3.45, p = 0.001. These
results suggest that the interactive teaching method effectively enhances
student performance in mathematics."
Chi-Squared Tests
Historical Context
1. Calculate all the expected frequencies i.e. Ei for all the values of i= 1, 2,
3,…………n.
2. Take difference of each observed frequency (Oi) and the corresponding
expected frequency (Ei) for each value of i i.e. find (Oi – Ei)
3. Square the difference for each value of i, i.e. calculate (Oi – Ei)2
4. Divide each square difference by corresponding expected frequency i.e.
calculate (Oi – Ei)2 / Ei for all the values of i = 1,2,3…….n.
5. Add all these quotients obtained in step 4, then X2 = Summation (Oi - Ei)
2 / Ei is the value of chi – square.
Calculation
● Step –1: Set up Null hypothesis H0: No association exists between the
Attributes Alternative hypothesis H1: An association exists between the
Attributes.
● Step – 4: Find the table value of x2 for level of significance and degree of
Freedom
1. Scale of Measurement:
○ Data should follow a continuous or ordinal scale, such as test scores
or other interval data.
2. Random Sampling:
○ Data should be drawn from a random, representative sample of the
population.
3. Normal Distribution:
○ When plotted, the data should ideally form a bell-shaped, or
normal distribution.
4. Homogeneity of Variance:
○ The variances within each group should be approximately equal,
meaning the spread or dispersion in data points is similar across
groups.
T-Distribution Tables
Types of T-Tests
Where,
○ mean 1 and mean 2=The average values of each of the sample sets
○ s(diff)=The standard deviation of the differences in the paired data
values
○ n=The sample size (the number of paired differences)
○ n−1=The degrees of freedom
○ The remaining two types belong to the independent t-tests. The
samples of these types are selected independent of each other.
Where,
○ mean 1 and mean 2=Average values of
each of the sample sets
○ var 1 and var 2=Variance of each of the sample sets
○ n1 and n2=Number of records in each sample set and,
○ Degrees of Freedom=n1+n2−2
where: n1 and n2=Number of records in each sample set.
● The unequal variance t-test is used when the number of samples in each
group is different, and the variance of the two data sets is also different.
● This test is also called Welch's t-test.
● Formula:
where:
○ mean 1 and mean 2=Average values of each
of the sample sets
○ var1 and var2=Variance of each of the sample sets
○ n1 and n2=Number of records in each sample set
and,
where:
○ var1 and var2=Variance of each of the sample sets
○ n1 and n2=Number of records in each sample set.
Applications of t-test
Purpose
Assumptions in ANOVA
ANOVA operates under several assumptions, which must be met to ensure the
validity of the test results:
1. Normality:
○ Data for each group should ideally follow a normal distribution.
○ This assumption is crucial for the reliability of ANOVA results,
especially in smaller samples.
2. Homogeneity of Variances:
○ Variances within each group should be approximately equal. This
assumption, known as homoscedasticity, means that the spread of
data within each group is similar.
○ When this assumption is violated, statistical techniques such as
Welch’s ANOVA may be more appropriate.
3. Independence of Observations:
○ Observations within each group must be independent of each other.
○ This assumption ensures that individual data points do not influence
one another, which is essential in experimental design.
4. Random Sampling:
○ Samples should be randomly drawn from the population, ensuring
that they are representative and unbiased.
1. Calculate Group Means: Find the mean of each group in the dataset.
2. Calculate the Overall Mean: Calculate the mean across all groups.
3. Determine Between-Group Variance (MST):
○ Measure the deviation of each group mean from the overall mean.
○ Multiply by the number of observations in each group, then average
these deviations.
4. Determine Within-Group Variance (MSE):
○ Measure the deviation of each observation from its group mean.
○ Square these deviations, sum them, and average for each group.
5. Compute F-Statistic: Divide the MST by the MSE.
● Interpretation:
○ High F-Value: Indicates that between-group variance is
substantially greater than within-group variance, suggesting
significant differences among group means.
○ Low F-Value: Implies that observed differences among means are
likely due to random variation, supporting the null hypothesis that
group means do not differ.
● Critical Values:
○ ANOVA relies on F-distribution tables to determine the critical value
for a given confidence level. If the computed F-statistic exceeds the
critical value, the null hypothesis is rejected, indicating significant
differences among means.
Types of ANOVA
1. One-Way ANOVA:
2. Two-Way ANOVA: