Statistical Tests For Comparative Studies
Statistical Tests For Comparative Studies
Parametric tests assume that the data follows a specific probability distribution, such
as the normal distribution. These tests often require that certain assumptions about the
data, such as independence and homogeneity of variance, are met in order to provide
accurate results.
Nonparametric tests do not assume a specific probability distribution for the data.
These tests are often used when the data does not meet the assumptions of
parametric tests or when the data is measured on a nominal or ordinal scale.
Homoscedasticity means that the variances of the dependent variable are similar across all levels of the
independent variable(s).
1. Visual inspection of scatterplots: One way to assess homoscedasticity is by visually examining scatterplots of the
dependent variable against the independent variable(s). If the spread of data points appears relatively equal
across the range of the independent variable(s), it suggests homoscedasticity. However, if the spread widens or
narrows as the independent variable(s) change, it indicates potential violations of homogeneity of variance.
2. Inspection of box plots or whisker plots: Box plots can provide a visual representation of the distribution and
spread of data across different groups or conditions. Check if the boxes are roughly equal in size or if the
whiskers have similar lengths across groups. Unequal box sizes or varying whisker lengths suggest potential
violations of homoscedasticity.
3. Levene's test: Levene's test is a statistical test used to formally assess the homogeneity of variance. It compares
the variances of the dependent variable between groups or conditions. The null hypothesis is that the variances
are equal. If the p-value associated with Levene's test is greater than the significance level (typically 0.05), it
suggests that the assumption of homogeneity of variance is met. However, if the p-value is below the significance
level, it indicates evidence against homoscedasticity.
4. Barlett's test: Barlett's test is another statistical test used to evaluate homogeneity of variance. It is similar to
Levene's test but is more sensitive to departures from normality. Barlett's test assumes that the data are normally
distributed. If the p-value from Barlett's test is greater than the significance level, it suggests that the assumption
of homogeneity of variance is met. Conversely, a p-value below the significance level indicates evidence against
homoscedasticity.
5. Residual plots: Residual plots are useful when assessing homoscedasticity in regression analysis. They involve
plotting the residuals (differences between observed and predicted values) against the predicted values. If the
spread of residuals remains relatively constant across the range of predicted values, it indicates
homoscedasticity. However, if the spread widens or narrows systematically as the predicted values change, it
suggests heteroscedasticity (violation of homogeneity of variance).
To determine the independence assumption for parametric tests, you need to assess whether the observations or
data points in your study are independent of each other. Independence implies that the values of the dependent variable
for one observation are not related to or influenced by the values of the dependent variable for other observations.
The technique for assessing independence depends on the design and nature of your data. Here are some common
approaches:
1. Study Design: If you have designed your study with a randomized or controlled design, such as a randomized
controlled trial (RCT) or a matched-pair design, the assumption of independence is often satisfied by the random
assignment or matching process.
2. Sampling Method: If you have collected data through random sampling, where each observation has an equal
chance of being selected, the assumption of independence is usually met. However, if your sampling method
involves a non-random process, such as convenience sampling or purposive sampling, the assumption of
independence may be violated.
3. Time or Order: For longitudinal or repeated measures data, it is essential to consider the temporal or sequential
order of the measurements. The assumption of independence may be violated if there is a systematic
relationship or dependency between the observations at different time points or measurement occasions.
4. Clustered or Nested Data: If your data is clustered or nested, such as participants nested within groups or
students nested within classrooms, the independence assumption may not hold. In such cases, you need to
account for the clustering or nesting structure in your analysis using appropriate statistical techniques (e.g.,
multilevel modeling).
5. Autocorrelation: Autocorrelation occurs when the values of the dependent variable for one observation are
correlated with the values of the dependent variable for neighboring observations. Autocorrelation violates the
independence assumption. Time series analysis or methods like autoregressive integrated moving average
(ARIMA) can be used to address autocorrelation in data that has a temporal component.
In comparative studies, the basis for rejecting the null hypothesis is typically the
statistical evidence obtained from data analysis. The statistical tests generate a test
statistic and a corresponding p-value. The general rule of thumb is that if the p-
value is below a predetermined significance level (commonly set at 0.05 or 0.01),
the null hypothesis is rejected.
The null hypothesis represents the absence of a difference, or effect between the
groups or conditions being compared. Rejecting the null hypothesis implies that the
observed data is unlikely to occur under the assumption of no effect or difference. It
suggests that there is sufficient evidence to support an alternative hypothesis, which
proposes the existence of a difference, or effect between the groups or conditions
being compared.