Untitled Document (44)
Untitled Document (44)
Non-parametric statistics are a set of statistical methods used to analyze data without assuming a specific
distribution (such as normal distribution) for the population from which the samples are drawn.
Advantages:
Flexible with fewer assumptions.
Applicable to a wider range of data types.
Robust against outliers and skewed data.
Disadvantages:
Generally less powerful than parametric tests (i.e., require larger samples to detect effects).
Provide less detailed information (e.g., do not estimate means or variances).
Definition Tests that assume the data follow a Tests that do not assume any specific
specific distribution, usually population distribution.
normal.
Data Type Require interval or ratio scale Suitable for nominal or ordinal data, and
data (continuous). also continuous data not meeting
parametric assumptions.
Sample Size Typically require larger sample Can be used with small sample sizes.
sizes for accuracy.
Test Use parameters like mean and Use ranks or frequency counts rather than
Statistics standard deviation. means.
Application When data meet assumptions and When data violate parametric assumptions
scale of measurement is or are ordinal/nominal.
appropriate.
2. Level of Measurement:-Can be used with nominal, ordinal, or interval/ratio data (especially when assumptions
for parametric tests are violated).
3. Independence- Observations should be independent of each other (except in some paired designs like Wilcoxon
signed-rank test).
4. Random Sampling- Samples should be drawn randomly from the population.
5. Shape of Distribution- Many non-parametric tests do not assume any specific shape for the distribution of
data.
6. Homogeneity of Variance:- Usually not required; non-parametric tests are robust against heterogeneity of
variance.
7. Sample Size:- Can be applied to small or large samples, but very small samples might reduce test power.
2. Use with Ordinal or Nominal Data:- Suitable for ordinal, nominal, or ranked data, as well as interval/ratio data
that violate parametric assumptions.
3. Based on Ranks or Signs:- Often use ranking or categorization of data rather than raw scores.
4. Robust to Outliers:- Less sensitive to outliers or extreme values in the data.
5. Less Restrictive Assumptions:- Require fewer assumptions about the data, making them more flexible.
6. Suitable for Small Sample Sizes:- Can be used effectively when sample sizes are too small to satisfy parametric
assumptions.
7. Generally Less Powerful:= Usually have less statistical power compared to parametric tests when assumptions
for parametric tests are met.
8. Simple Computation:- Often simpler to compute, especially before the widespread use of computers.
Spearman’s Rank Difference Correlation
Spearman’s Rank Difference Correlation is a non-parametric measure that assesses the strength and direction of the
monotonic relationship between two variables based on the ranks of their values rather than the actual data. It uses the
differences between the ranks of paired observations to calculate the correlation coefficient.
2. Equation:
Where:
3. Steps to Calculate:
5. Advantages:
Does not require assumptions of normality or linearity.
Suitable for ordinal data or continuous data that is not normally distributed.
Robust to outliers as it uses ranks instead of raw data.
Can detect monotonic relationships (either increasing or decreasing).
6. Disadvantages:
Less powerful than Pearson’s correlation when data is truly interval and linear.
Cannot capture non-monotonic relationships.
Ties in ranks complicate calculation and interpretation (though correction methods exist).
Sensitive to sample size; very small samples may give misleading results.
7. Limitations:
Not suitable for nominal data (only ordinal or higher).
Assumes a monotonic relationship; cannot detect more complex relationships.
Presence of many ties can reduce accuracy.
Cannot specify causality—only association.
8. Alternatives:
Alternative Method When to Use
9. Scenario Example:
Imagine a researcher wants to study if students’ ranks in mathematics predict their ranks in physics. Since the interest is in
ranks and the data may not be normally distributed, Spearman’s rank difference correlation is appropriate. After ranking
both sets of scores, the researcher calculates the rank differences and applies the formula to find the correlation, which
indicates how closely related the performances are across subjects.
Biserial Correlation
Biserial correlation is a statistical measure used to determine the strength and direction of the association between:
● Point-biserial correlation: Used when the dichotomous variable is natural (truly categorical, e.g., male/female).
Biserial correlation: Used when the dichotomous variable is artificial (created by splitting a continuous variable
into two groups, e.g., pass/fail based on a cutoff score).
4. Assumptions:
● The dichotomous variable is artificially created by cutting a continuous variable at some point.
The underlying continuous variable is normally distributed.
Both groups have equal variances (homoscedasticity).
The relationship is linear.
5. Advantages:
6. Disadvantages:
● Requires the assumption of an underlying continuous normal distribution which may not always hold.
Calculation is more complex because of the need to estimate yy from the normal distribution.
If the dichotomy is natural, point-biserial correlation is more appropriate.
Sensitive to the cutoff point chosen to dichotomize the continuous variable.
7. Example Scenario:
Suppose an educational researcher has a continuous variable: students’ intelligence scores. They create two groups by
dividing students into "high IQ" and "low IQ" based on a cutoff score of 100.
● The researcher wants to see how this artificial grouping correlates with another continuous variable such as
academic performance.
Using biserial correlation, the researcher can estimate the strength of association between the underlying
continuous IQ distribution and academic performance more accurately than using point-biserial correlation.
Point-Biserial Correlation
Point-biserial correlation is a statistical measure used to calculate the strength and direction of the relationship between:
2. Formula:
Where:
3. Assumptions:
The dichotomous variable is naturally occurring (not artificially created).
The continuous variable is normally distributed within each group.
Groups have homogeneity of variance (equal variances).
The relationship is linear.
4. Advantages:
Simple to compute and interpret.
Measures association between binary and continuous variables effectively.
Widely used in psychological, educational, and social research.
Directly related to t-test for difference between two means.
5. Disadvantages:
Assumes natural dichotomy; inappropriate if the binary variable is artificially created.
Sensitive to unequal group sizes (unbalanced groups).
Requires normality and equal variance assumptions.
Does not measure causality, only association.
6. Example Scenario:
A researcher studies whether there is a relationship between gender (male = 0, female = 1) and math test scores
(continuous variable).
Tetrachoric Correlation
Tetrachoric correlation is a measure of association between two dichotomous variables that are assumed to represent
underlying continuous latent variables which have been artificially dichotomized.In other words, it estimates the
correlation between two unobserved continuous variables from the observed binary data.
2. When to Use:
Variables Two observed binary variables Two latent continuous variables behind binaries
Assumptions No assumptions about latent traits Assumes underlying bivariate normal distribution
● The formula is complex and usually computed via iterative methods or software.
Based on the frequencies in a 2x2 contingency table, and the thresholds for dichotomizing the latent continuous
variables.
Uses the bivariate normal distribution to estimate correlation.
5. Assumptions:
Each observed binary variable represents an underlying normally distributed continuous variable.
The latent continuous variables are linearly related.
The dichotomization thresholds are known or estimated.
The sample size is adequate for stable estimation.
6. Advantages:
Provides more accurate estimates of association for dichotomous variables from continuous origins.
Useful in psychometrics when testing relationships between binary test items.
Corrects downward bias of simpler measures like phi coefficient when artificial dichotomization exists.
7. Disadvantages:
Computationally intensive (requires special software like R, SPSS with add-ons, or dedicated packages).
Assumptions (normality of latent variables, linearity) may not always hold.
Not suitable for naturally dichotomous variables (use point-biserial or phi instead).
Sensitive to small sample sizes.
8. Example Scenario:
A researcher studies the correlation between two dichotomous test items (pass/fail):
9. Interpretation:
the value of tetrachoric correlation ranges from -1 to +1.
A high positive value suggests a strong positive relationship between the latent continuous traits.
Unlike phi, it corrects for the artificial dichotomization.
10. Applications:
Psychometric test analysis (item analysis).
Behavioral sciences when dealing with dichotomized continuous data.
Structural equation modeling when latent variables are involved.
It is equivalent to the Pearson correlation coefficient when both variables are dichotomous.
2. When to Use:
3. Formula:
4. Range
6. Advantages:
Sensitive to marginal totals; can be misleading if marginal frequencies are very uneven.
Does not account for underlying latent continuous variables if the binary variables come from dichotomizing
continuous variables.
Limited to 2x2 tables (only two dichotomous variables).
8. Example:
9. Applications:
. Related Measures:
Use Tetrachoric correlation when variables are dichotomized continuous variables with latent normality assumptions.
Use Point biserial correlation when one variable is dichotomous and the other is continuous.
It tests if the observed frequencies in categories differ significantly from the expected frequencies under the null
hypothesis of independence.
Chi-Square Test of Tests association between two categorical variables Two categorical variables
Independence
Chi-Square Goodness of Fit Tests if observed frequency distribution fits an One categorical variable
Test expected distribution
3-formula
7. Advantages:
Simple and widely used test for categorical data.
No assumptions about normality or distribution shape (non-parametric).
Applicable for nominal and ordinal data.
Tests independence without requiring interval data.
8. Disadvantages:
Requires sufficiently large sample size for accuracy.
Sensitive to small expected frequencies; can give misleading results if expected counts <5.
Does not indicate strength or direction of association, only existence.
Cannot be used for paired or matched data without modifications.
9. Example: A researcher studies the association between Gender (Male/Female) and Preference for Tea or Coffee
among 100 people(chatgpt)
10. Applications:
Mann-Whitney U Test
The Mann-Whitney U Test (also called the Wilcoxon rank-sum test) is a non-parametric test used to compare
differences between two independent groups on an ordinal or continuous variable when the assumptions of the
parametric independent t-test are not met (e.g., non-normal distribution).
. When to Use:
Comparing two independent samples/groups.
Data are ordinal or continuous but not normally distributed.
Sample sizes can be small or unequal.
Data do not require interval scaling.
3. Assumptions:
Observations are independent between groups.
The dependent variable is at least ordinal.
The shapes of the distributions of the two groups are similar (for comparing medians).
Samples are randomly selected.
4. Hypotheses:-Null hypothesis (H₀): The two populations have the same distribution (no difference in ranks).
Alternative hypothesis (H₁): The two populations differ in distribution.
6. Steps to Perform Mann-Whitney U Test:
1. Combine data from both groups and rank all values from lowest to highest.
Assign tied ranks the average of their positions.
Calculate the sum of ranks (R1R_1, R2R_2) for each group.
Compute U statistics U1U_1 and U2U_2.
Find the smaller U value.
Compare U to critical values from Mann-Whitney U tables (based on n1n_1 and n2n_2) or calculate p-value (with
normal approximation for large samples).
Decide to reject or fail to reject H0H_0.
7. Advantages:
Does not require normal distribution.
Suitable for small samples and ordinal data.
Can handle unequal sample sizes.
Robust to outliers.
8. Disadvantages:
Less powerful than parametric t-test if data are normally distributed.
Tests differences in distributions, not specifically means.
Interpretation focuses on ranks, not raw data.
Assumes similar distribution shape to compare medians meaningfully.
9. Example:
Suppose we want to compare stress scores (measured on an ordinal scale) between two independent groups:(chatgpt)
2. When to Use:
Comparing two related or paired samples (e.g., pre-test/post-test scores).
Data are ordinal or continuous but not normally distributed.
Sample size is small or data violate parametric assumptions.
3. Assumptions:
The data are paired and come from the same subjects or matched pairs.
Differences between pairs are independent.
The dependent variable is at least ordinal.
The distribution of differences is symmetric (less strict than normality).
4. Hypotheses:
● Null hypothesis (H₀): The median difference between pairs is zero (no change).
Alternative hypothesis (H₁): The median difference is not zero (there is change).
5. Test Procedure:
7. Advantages:
Does not require normal distribution.
Suitable for small sample sizes and ordinal data.
Uses magnitude and direction of differences.
More powerful than the sign test as it uses rank information.
8. Disadvantages:
Less powerful than the parametric paired t-test if normality holds.
Assumes differences are symmetric in distribution.
Cannot be used if many zero differences occur.
Interpretation is based on ranks, not raw scores.
9. Example:
Suppose a researcher measures anxiety levels before and after therapy in 6 patients(chatgpt)
10. Applications:
Kruskal-Wallis H Test
The Kruskal-Wallis H Test is a non-parametric test used to compare three or more independent groups to determine
if there are statistically significant differences between their median ranks. It is the non-parametric alternative to the
one-way ANOVA when the assumptions of normality and/or homogeneity of variance are violated.
2. When to Use:
4. Hypotheses:
● Null hypothesis (H₀): The populations have the same median (no difference among groups).
Alternative hypothesis (H₁): At least one group differs in median from others.
5. Test Procedure:
1. Combine all data from all groups and rank all observations together from lowest to highest, ignoring group
membership.
Calculate the sum of ranks for each group.
Compute the Kruskal-Wallis test statistic HH using the formula:
2. Where:
7. Disadvantages:
Less powerful than parametric ANOVA if assumptions for ANOVA hold.
If significant, does not specify which groups differ (post hoc tests needed).
Assumes groups have similar shaped distributions.
Based on ranks, so loses some information about magnitude.
8. Example:
A researcher wants to compare stress levels among three groups of employees working in different departments (A, B, C).
Stress scores are ordinal and non-normal:
8 10 7
9 11 6
7 13 5
●
Combine all scores, rank them from 1 (lowest) to 9 (highest).
● Sum ranks for each group and calculate HH.
● Check if stress levels differ across departments.
Median Test
The Median Test is a non-parametric test used to determine if there are differences in the medians of two or more
independent groups. It tests whether the groups come from populations with the same median.
2. When to Use:
Comparing the medians of two or more independent groups.
Data are ordinal or continuous but do not meet parametric test assumptions (e.g., non-normal distribution).
Suitable for small samples or skewed data.
3. Assumptions:
Observations are independent within and between groups.
The measurement scale is at least ordinal.
Groups are independent.
4. Hypotheses:
5. Test Procedure:
6. Advantages:
Less sensitive (less powerful) than other non-parametric tests like Kruskal-Wallis.
Only compares medians, ignoring other distribution features.
Reduces data to above/below median, losing information.
8. Example:
Researchers want to test if three diet plans differ in their median weight loss. They record weight loss for participants and:
Calculate the overall median weight loss.
Count how many participants in each diet group lost more or less than the median.
Perform a Chi-square test on these counts.
9. Applications: