Midterm II Review: Remember: Include Lots of Examples, Be Concise But Give Lots of Information
Midterm II Review: Remember: Include Lots of Examples, Be Concise But Give Lots of Information
1.
Differentiate between and within subjects design
o A between subjects design is when you have separate groups that undergo different treatments (ie.
Group 1 goes through condition A, group 2 condition B). A within subjects design is when there is
only one group of participants and they undergo ALL of the experimental procedures (ie the group
receives condition A, B, and C).
Know assumptions of ANOVA
o Normality
Normal distribution of data, hard to test especially with n < 30. It is a theoretical assumption.
Kolmogorov-Smirnov Test tests for normalcy, but is only valid for large n.
o Homogeneity
Explained Error
Unexplained
Error
o MSb = SSbetween/dFbetween
Reflects variability among group means.
o MSw = SSwithin/dFwithin
Reflects variance within a group, aka variability among scores within a sample. This is bad,
and we want to control for this. Increased sample size leads to decrease in variance within a
group. An increase in within mean squared error, it could mask a significant effect. Also
equal to the squared standard error
o F = MSbetween/MSwithin
F is a ratio between the mean squared error between and within. An increased F means the
difference is between the groups (good) while a low F means most of the variability is due to
difference within the groups (variability among subjects) which is bad. In a simple
comparison, F = t2.
If the null hypothesis is true, the within group error estimate will be the same as the between
group error estimate (1:1). If the alternative hypothesis is true (we reject the null) there is a
significant difference between the variance between groups and the variance within groups.
This means the variance is caused by differences within our treatment groups (good).
Variation within Variation between
population population
Null is true
Within group estimate X
of variance
Between group X
estimate of variance
Null is false
Within group estimate X
of variance
Between group X X
estimate of variance
Why use F? We use F for hypothesis testing – we calculate the statistic, and compare the stat
to the sampling distribution of means for that statistic. They have 2 different Degrees of
freedom (dFw and dFb).
Know 3 effect sizes
o Effect size reflects the proportion of variance accounted for by the measure. Increased 2 is good
because it means most of the variance is due to difference between the groups. With one level of
variability, the values are the same. With more complex designs (2 levels or more) partial is better
(more accurate). The estimates will differ. Effect size is just the Correlation coefficient.
o Eta squared (2)
R2 = SSbetween/SStotal
o Partial Eta (partial 2)
R2 = SSbetween / (SSbetween+SSerror(within))
More typically reported than eta squared.
o Generalized (generalized 2)
Standardized test used for Repeated Measures ANOVAs
3.
Define Factorial ANOVA
An ANOVA performed with 2 or more groups.
SS Df in MS R2 F
Factorial
ANOVA
Rows SSrows Nrows – 1 SSrows/dFrows SSR / (SStot – SSC MSrows / MSwithin
– SSint)
Columns SScolumns Ncolumns – 1 SScolumns/dFcolumns SSC / (SStot – SSR MScolumns /
– SSint) MSwithin
Interaction SSinteraction N cells – df SSinteraction/dFinteraction SSint / (SStot – SSR MSinteraction/
rows – df – SSC) MSwithin
columns - 1
Within df1 + df 2 + SSwithin/dFwithin
(x-xbar)2
df 3 + … +
dfn (need to
know #
subjects in
groups.)
Total N-1
(x-xGM)2
SSb + SSw
Know Main Effect (ME) and Interactions
o Main effect is the effect of an independent variable on a dependent variable averaging across the
levels of any other independent variables. “a” shows main effect of group, “b” shows the main effect
of the group. The lines are parallel.
o Interaction is when the relationship among three or more variables, and describes a situation in which
the simultaneous influence of two variables on a third is not additive. The lines are not parallel.
Recognizing main effects and interactions
o When you run analysis you are coming up with model/theory that explains your data. Ie correlation
coefficient: how well your line fits. 1 is perfect. ANOVA is essentially just this, and effect size is just
your correlation coefficient (R2). Be careful when judging main effect or interaction – outliers may
throw you off. For example, there may appear to be a ME, but that ME may be driven by interaction.
Variance estimates in factorial design
Within group variance estimate – as before, the within group variance estimate reflects the average of the
population variance estimates made from the scores for each cell.
sw2= (s12 + s22 + s32 + … + sn2) / K *****
MSwithin
group
condition s22 s32
s42 s52
SSwithin/dFwithin
Main effect variance estimate - as in single level design, the main effect between variance estimate is
based on the variation between the column / row means. This is like doing 2 one way ANOVAs, one for
condition effect and one for group effect. Get a p value for both.
Interaction variance estimate – interaction variance estimated is based on the variations between the other
possible cell groupings. Receive a p value for this as well.
Unexplained
variance
ME 1
ME 2
Interaction
Unexplained error (error within) is the deviance of your scores from the mean of the group, or the
variance of the group. In a perfect world, there would be no error within, and there would be variance
between the group means (MSbetween). We need ANOVA to detect if there is significant variance
between groups. Small samples are bad. Outliers increase MSw (bad). In a factorial ANOVA, the MSw is
calculated
R2 – proportion of explained variance. If SSe is big, all the between R2 will be small. A big R2 (i.e. rows)
means variance was mostly driven by that R2. If the effect size is small, it means it was a bad experiment
(black t-shirt example). Literally explains how the variance is accounted for.
POSTHOC analysis of factorial ANOVA
Tukey’s output would create a large number of tests: the more tests done, the more likely you are to
increase the type I error. You can correct for this using the Bonferroni correction.
POSTHOC with complex design is generally hard. Consider:
Plotting data
Simple effects analysis (t-tests) (perform these after plotting data, make smart comparisons)
4.
Repeated Measures ANOVA
o All subjects are exposed to all conditions, OR repeating the same test across time with the same
individuals.
Know how its variance differs from other ANOVA
o Partitioning variance differently – subtract subject variance from unexplained error.
Explained Error
Unexplained
Error
Subject error
o SStotal = SS conditions (deviation of condition means from grand mean. Same as SSbetween in
regular ANOVA) + SS subject + SS error (deviation of subject scores from group mean)
o SS subject (subject error) = deviation of subject means from grand mean.
Explain Sphericity
o Sphericity is when the variance of difference scores is equal. When Compound Symmetry is met,
Sphericity is met. It is essential to have sphericity or the ANOVA does not work.
o Dealing with violations of sphericity:
Maulchy’s Test. Use if n > 30. If Maulchy’s test is significant, run Greenhouse-Geisser.
Use Greenhouse-Geisser correction whenever n < 30.
POST-HOC with paired-samples T-test. If you have lots of columns, it is possible to use
Bonferroni correction.
Tutorials / Discoveries.