PARAMETRIC TESTS pdf
PARAMETRIC TESTS pdf
Parametric tests are powerful statistical tools, but their validity hinges
on a set of critical assumptions. If these assumptions are met,
parametric tests can provide accurate, reliable, and generalizable
results. However, when these conditions are violated, the conclusions
drawn from the analysis may be misleading or incorrect. Therefore,
understanding the assumptions of parametric tests is essential for
every researcher or analyst.
1. Normality of Data
Levene’s Test
Bartlett’s Test
Brown-Forsythe Test
3. Independence of Observations
The assumption of independence means that the data points should
not influence each other. Each subject or observation must be
independent from the others. This is a foundational requirement
across nearly all parametric tests.
Violations of this assumption often occur in:
Parametric tests are suitable for interval and ratio scale data, where
distances between values are meaningful and consistent. Examples
include income, temperature (in Celsius or Fahrenheit), height, and
weight.
Scatterplots
Residual plots
Polynomial regression (to explore curvilinear relationships).
Implications of Violating Assumptions
To mitigate this:
One-sample t-test
Independent (two-sample) t-test
Paired (dependent) t-test
Assumptions:
Formula:
Where:
xˉ\bar{x}xˉ = sample mean
μ\muμ = population mean
sss = sample standard deviation
nnn = sample size
Example Scenario:
A researcher compares the average test scores of two different
classrooms using different teaching methods.
Assumptions:
t=xˉ1−xˉ2sp2(1n1+1n2)t = \frac{\bar{x}_1 -
\bar{x}_2}{\sqrt{s_p^2(\frac{1}{n_1} + \frac{1}{n_2})}}t=sp2(n11
+n21)xˉ1−xˉ2
The paired t-test is used when the same group is tested twice (pre-test
and post-test), or when participants are matched in pairs (e.g., twins or
spouses). It evaluates the mean of the differences between paired
observations.
Example Scenario:
A psychologist wants to test the effectiveness of a therapy by
measuring anxiety levels before and after treatment in the same
patients.
Assumptions:
Applications of t-Tests
This test is used when there is one independent variable with three or
more levels (groups) and one continuous dependent variable.
Example Scenario:
A teacher wants to compare the exam scores of students from three
different teaching methods to determine which method is most
effective.
Assumptions:
Hypotheses:
Example Scenario:
A researcher studies the impact of study method (visual/audio/text)
and study duration (1 hour/2 hours) on student test performance.
Main Effects:
Interaction Effect:
Advantages:
Example Scenario:
A fitness trainer records the performance of athletes at three time
points: before training, mid-training, and post-training.
Advantages:
Assumptions:
Applications of ANOVA
Strengths:
Limitations:
Assumes equal variances and normality.
Can become complex with multiple factors and interactions.
Doesn’t identify which means differ without post hoc testing.
Example Scenario:
A researcher wants to predict a student’s final exam score based on
their number of study hours.
Where:
Interpretation:
The slope (β1\beta_1β1) indicates how much YYY changes for
a one-unit increase in XXX.
The intercept (β0\beta_0β0) represents the expected value of
YYY when X=0X = 0X=0.
Assumptions:
Linearity between XXX and YYY
Independence of observations
Homoscedasticity (constant variance of residuals)
Normality of residuals
Output includes:
R-squared: proportion of variance in YYY explained by XXX
Coefficients table (with p-values)
Residual plots (to check assumptions)
Equation:
Benefits:
Captures complex, real-world relationships.
Allows for control of confounding variables.
Enables prediction and forecasting.
Linearity
Independence
Homoscedasticity
Normality of residuals
No multicollinearity (predictors should not be highly correlated)
Strengths:
Limitations:
Sensitive to outliers.
Misleading results if assumptions are violated.
Overfitting when too many predictors are used.
3.4 Z-Test
The z-test is one of the simplest parametric statistical tests, used to
determine whether there is a significant difference between sample
and population means or between two sample means. It is similar to
the t-test but is used when the population standard deviation is known
and the sample size is large (typically n > 30).
One-sample z-test
Two-sample z-test
Z-test for proportions
Example Scenario:
A manufacturing company claims their light bulbs last 1,200 hours. A
quality control analyst tests a sample of bulbs and uses a one-sample
z-test to determine if the claim holds.
Formula:
Where:
xˉ\bar{x}xˉ = sample mean
μ\muμ = population mean
σ\sigmaσ = population standard deviation
nnn = sample size
This test compares the means of two independent samples when both
population standard deviations are known and the sample sizes are
large.
Example Scenario:
Comparing the average heights of male and female employees at a
company using large samples.
Formula:
z=(xˉ1−xˉ2)(σ12/n1)+(σ22/n2)z = \frac{(\bar{x}_1 -
\bar{x}_2)}{\sqrt{(\sigma_1^2/n_1) + (\sigma_2^2/n_2)}}z=(σ12/n1
)+(σ22/n2)(xˉ1−xˉ2)
Example Scenario:
A political analyst compares approval ratings between two candidates
using sample proportions.
Formula:
z=p1−p2p(1−p)(1n1+1n2)z = \frac{p_1 - p_2}{\sqrt{p(1 -
p)(\frac{1}{n_1} + \frac{1}{n_2})}}z=p(1−p)(n11+n21)p1−p2
Applications of Z-Test
Limitations
Parametric tests are widely regarded as the gold standard in statistical inference when their
assumptions are met. Below are some key advantages:
Parametric tests are more statistically powerful than their non-parametric counterparts,
meaning they are more likely to detect a true effect or difference when one exists. This power
comes from the fact that they use more information from the data, such as the mean and
standard deviation.
Example:
A t-test will usually detect smaller differences between group means more efficiently than a
non-parametric test like the Mann–Whitney U test.
When data follows a normal distribution, parametric tests offer precise and efficient results.
They use the actual values and rely on probability theory, making the outcomes more reliable
and robust under the right conditions.
Parametric tests like ANOVA and regression analysis can handle multiple independent
variables, interaction effects, and covariates, which is often not possible with non-parametric
tests.
Example:
A Two-Way ANOVA can test the interaction between teaching style and student gender,
something that non-parametric tests typically can't handle without complex adaptations.
Parametric tests are well-supported in statistical software packages such as SPSS, R, Python
(SciPy/Statsmodels), and Excel. These tools offer built-in functions, confidence intervals,
effect sizes, and visualizations.
The results of parametric tests are often easier to interpret and have standard reporting
formats in academic writing. For example, reporting a regression coefficient or an F-value in
ANOVA is universally understood and accepted in scholarly communities.
Despite their advantages, parametric tests are not always suitable and have notable
limitations.
1. Strict Assumptions
Parametric tests rely on a number of strict assumptions:
Example:
Using a t-test on skewed data with outliers can produce misleading p-values.
2. Sensitivity to Outliers
Parametric tests are sensitive to outliers, which can significantly distort the results because
they affect both the mean and the standard deviation.
Example:
In regression analysis, a few extreme values can change the slope of the line and reduce the
model's validity.
Parametric tests are generally not suitable for ordinal or nominal data (e.g., survey rankings,
yes/no responses). Such data violate the assumption of equal intervals, making non-
parametric alternatives like Chi-Square or Kruskal-Wallis more appropriate.
Although parametric tests can be used with small samples, their accuracy diminishes if the
normality assumption cannot be verified. In such cases, non-parametric tests are safer.
When assumptions are not met, analysts must perform additional steps like data
transformation, using robust statistical techniques, or switching to non-parametric tests—
adding complexity to the analysis.
Conclusion
Parametric tests form the backbone of modern statistical analysis,
offering a robust and efficient framework for making inferences about
populations based on sample data. Rooted in probability theory and
the properties of the normal distribution, these tests are widely
applicable across disciplines including finance, medicine, education,
psychology, and the natural sciences.