CSS
CSS
1.Independent t-test:
2. Paired t-test:
- Purpose: Compare means within the same group at two time points (e.g., pre- and post-test).
- SPSS Steps: Analyze > Compare Means > Paired Samples T-test.
- Report:
- df, t-value, p-value.
- Effect Size: Use Cohen’s d for paired samples: d = (Mean difference) / SD difference.
- Purpose: Predict the value of a dependent variable based on one or more independent
variables.
- SPSS Steps: Analyze > Regression > Linear.
- Check Descriptives: Mean, standard deviation, etc.
-Residual stats: CHECK MIN/MAX. THE
- R² (R2/R2 adjusted!): Indicates the proportion of variance in the DV explained by the IV(s).
Report R² to 2 decimal places.
- Durbin-Watson: Test for autocorrelation (0-4 range; close to 2 is ideal IF IT IS –
ASSUMPTION OF RESIDUALS ISN’T VIOLATED).
- ANOVA Table: Check significance of F-statistic to see if the model fits better than baseline (p
< .05).
- df1 = Number of predictors; df2 = N - df1 - 1.
- CHECK COEFFICIENTS TABLE FOR SIGNIFICANCE OF VARIABLES, RELATIONSHIPS:
- Unstandardized B: Represents change in DV for each unit change in IV.
- Standardized Beta (β): For comparing relative strength of predictors.
- Significance (p-value): p < .05 means predictor is significant.
MULTICOLLIN (ASSUMPTION)
IMPORTANT
FIRST/SECOND MODEL AND WHETHER THEY ARE SIGNIFICANTLY BETTER THAN THE
BASELINE MODEL (SIG. F CHANGE)
5. Logistic Regression:
DIRECTION OF RELETAIONSHIP
REPORT B, EXP (B), WALD, CI
- Model Summary: Report -2 Log Likelihood, Cox & Snell R², and Nagelkerke R² (pseudo R²).
- Exp(B): Report odds ratios for each predictor (Exp(B)).
Key Points About Durbin-Watson:
Interpreting Durbin-Watson:
DW ≈ 2: Residuals are independent; no autocorrelation is detected.
DW < 2: Indicates positive autocorrelation in the residuals.
DW > 2: Indicates negative autocorrelation in the residuals.
RESIDUALS
Typical Thresholds
Standardized residuals typically follow a normal distribution:
o Values close to 0 indicate that the observation is close to the predicted value.
o Values greater than +2 or less than -2 may indicate outliers, and values greater
than +3 or less than -3 are often considered highly unusual.
Conclusion
In summary, minimum and maximum standardized residuals provide insights into the
behavior of individual data points in relation to your regression model:
o Minimum: Highlights observations that are significantly lower than predicted.
o Maximum: Highlights observations that are significantly higher than predicted.
ASUMPTIONS
Linear Regression
Key Assumptions:
Linearity: The relationship between the independent and dependent variables is linear.
Independence: The residuals (errors) are independent (no autocorrelation). Checked
using the Durbin-Watson test.
Homoscedasticity: The residuals have constant variance across all levels of the
independent variables (checked with a scatterplot of residuals vs. predicted values).
Normality of Residuals: The residuals should be approximately normally distributed.
Checked using histograms, normal probability plots, or the Shapiro-Wilk test.
No Multicollinearity: The independent variables are not too highly correlated. Checked
using VIF (Variance Inflation Factor).
Logistic Regression
Key Assumptions:
Linearity of Logits: The logit transformation of the dependent variable should have a
linear relationship with the independent variables.
Independence: The observations are independent.
No Multicollinearity: The independent variables should not be highly correlated.
Checked using VIF.
Adequate Sample Size: Logistic regression requires a larger sample size than linear
regression.
3. No Multicollinearity:
o In the logistic regression dialog (Analyze > Regression > Binary Logistic), click
Save.
o Check the box for Collinearity diagnostics.
o After running the regression, check VIF values under Coefficients (available if
collinearity is requested). Like linear regression, VIF should be below 5.
4. Adequate Sample Size:
o Make sure you have enough cases, especially for each level of your categorical
variables. SPSS does not automatically test this, so ensure your sample size is
large enough to detect meaningful results (at least 10 cases per predictor).
MULTICOLINEARITY LOGISTIC
4. Example of Interpretation:
Imagine you’re predicting whether someone will pass or fail a test (outcome: pass = 1, fail = 0).
The predictors are study time and age.
If study time has Exp(B) = 2.0, you’d say that for each additional hour of study, the
odds of passing the test are doubled.
If age has Exp(B) = 0.9, you’d say that for each additional year of age, the odds of
passing the test decrease by 10%.
In summary:
Positive coefficients (B) or Exp(B) greater than 1: Increase in likelihood of the outcome.
Negative coefficients (B) or Exp(B) less than 1: Decrease in likelihood of the outcome.
Assumption Checks:
1. Normality:
- Purpose: To verify if your data is normally distributed, a key assumption for parametric tests
like t-tests and ANOVA.
Steps in SPSS:
- Analyze > Descriptive Statistics > Explore.
- Move the variable(s) to the Dependent List.
- Click Plots and check Normality plots with tests.
Interpretation:
- If p > .05: The data is normally distributed.
- If p < .05: Data deviates significantly from normal distribution.
Visual Check:
- Use Q-Q Plot or Histogram (also available under Plots) to visually assess normality.
- Q-Q Plot: Points should fall along the diagonal line if normal
2. Multicollinearity:
- Purpose: To ensure your independent variables are not too highly correlated (important for
regression models).
Interpretation:
- Variance Inflation Factor (VIF):
- VIF < 5: Acceptable (closer to 1 is ideal).
- VIF > 10: Indicates problematic multicollinearity.
- Tolerance:
- Tolerance > .2: Acceptable.
- Tolerance < .2: Suggests multicollinearity.
3. Linearity:
- Purpose: Assumes a linear relationship between dependent and independent variables in
regression analysis.
SPSS Linearity:
- Analyze > Regression > Linear.
- Click on Plots.
- Place ZRESID (Standardized Residuals) on the Y-axis and ZPRED (Standardized Predicted
Values) on the X-axis.
- Click Continue and OK.
Interpretation:
- Look at the scatterplot of residuals vs. predicted values. If the points are randomly dispersed,
the assumption of linearity is met.
- If there’s a clear pattern (e.g., curved), the linearity assumption may be violated.
4. Homoscedasticity:
- Purpose: Assumes that the residuals have constant variance at each level of the independent
variable(s) in regression.
Interpretation:
- If the spread of residuals is roughly the same across all values of predicted scores (random
cloud), homoscedasticity is met.
- If you see a funnel shape or other pattern, heteroscedasticity might be an issue (variance is
not constant).
---
5. Outliers:
- Purpose: Detect if extreme values are unduly influencing your results.
Interpretation:
- Look at Cook’s Distance in your output.
- Cook’s Distance > 1: Indicates an influential outlier.
Interpretation:
- Durbin-Watson (independence of observations) value should be between 1 and 3.
- Values close to 2 indicate no autocorrelation.
- Values close to 0 or 4 suggest positive or negative autocorrelation, respectively.
7. Effect Size:
- For t-tests:
- Cohen’s d: Small (0.2), Medium (0.5), Large (0.8).
- For correlation: r-value is the effect size.
- For regression: Use R² as effect size (proportion of variance explained).
9. SPSS Tips:
- Always check Descriptives before running tests.
- Look at Normality (histograms, Q-Q plots, Shapiro-Wilk) to check if assumptions hold.
- Use Transform > Compute to create new variables (e.g., log transformations).
- Use Plots for assumptions testing (Residuals vs Predicted, P-P plots).
ASSUMPTION VIOLATION (LINEAR)
To check whether the assumptions of linear regression are violated in SPSS, follow these steps
for each assumption:
1. Linearity
- What to Check: The relationship between the independent variables and the dependent
variable should be linear.
- How to Check:
- Run a linear regression analysis:
- Go to Analyze > Regression > Linear.
- Include your dependent variable and independent variables.
- In the Plots menu, select the Standardized Residuals for the Y-axis and Standardized
Predicted Values for the X-axis.
- Look for a random scatter of points around zero; if the plot shows a clear curve, this indicates
a violation of linearity.
2. Independence of Residuals
- What to Check: Residuals (errors) should be independent (no autocorrelation).
- How to Check:
- Conduct the Durbin-Watson test:
- After running your linear regression, check the Model Summary output for the Durbin-
Watson statistic.
- A value close to 2 suggests independence, while values below 1 or above 3 indicate
possible autocorrelation.
3. Homoscedasticity
- What to Check: The residuals should have constant variance at all levels of the independent
variables.
- How to Check:
- In the Plots menu of the linear regression dialog, select ZRESID (standardized residuals) for
the Y-axis and ZPRED (standardized predicted values) for the X-axis.
- Look for a random scatter of points without a distinct pattern. A funnel shape (wider spread of
residuals at higher levels of predicted values) indicates heteroscedasticity.
4. Normality of Residuals
- What to Check: The residuals should be approximately normally distributed.
- How to Check:
- Generate a Histogram of the residuals:
- Save the residuals in the regression output by going to Save in the linear regression dialog
and checking the box for Unstandardized Residuals.
- Then go to Graphs > Legacy Dialogs > Histogram and plot the histogram of the residuals.
- Additionally, you can run the Shapiro-Wilk test:
- Go to Analyze > Descriptive Statistics > Explore.
- Move your residuals variable into the Dependent List and check the option for Normality
plots with tests.
- If the Shapiro-Wilk test shows a significant p-value (typically p < 0.05), this indicates a
violation of normality.
5. No Multicollinearity
- What to Check: The independent variables should not be too highly correlated with each other.
- How to Check:
- In the linear regression dialog, click on Statistics, and check the box for Collinearity
Diagnostics.
- After running the regression, check the Collinearity Statistics in the output.
- Look for the Variance Inflation Factor (VIF) values:
- A VIF greater than 10 indicates high multicollinearity. As a rule of thumb, VIF values
between 1 and 5 are acceptable.
Steps to Report Unstandardized Coefficients (B)
A multiple linear regression analysis was conducted to predict the perceived profile picture
rating (FB_Profile_TOT) based on gender, age, extraversion, and narcissism. The results of the
regression analysis are summarized in the table below:
The unstandardized coefficient for the constant (B = 1.25) indicates that when all predictors are
at their reference levels, the average FB_Profile_TOT rating is expected to be 1.25. For every
one-unit increase in age, the FB_Profile_TOT rating increases by 0.10 (p < 0.01), suggesting
that older individuals tend to receive higher ratings. Additionally, for each unit increase in
extraversion, the rating increases by 0.30 (p < 0.01), highlighting the positive impact of
extraversion on perceived profile picture ratings. Narcissism also positively influences the rating,
with a coefficient of 0.20 (p = 0.02), although its effect is less pronounced than that of
extraversion. The coefficient for gender (B = 0.45, p = 0.04) suggests that males tend to receive
higher ratings compared to females.