0% found this document useful (0 votes)
11 views

Field Analysis in Stats

The document is the table of contents for the book 'Discovering Statistics Using IBM SPSS Statistics, 4th Edition' by Andy Field. It outlines various chapters covering topics such as the research process, statistical models, data analysis, and the use of SPSS software for statistical analysis. Each chapter includes sections on theory, practical applications, and further reading.

Uploaded by

nonyethegirl
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
11 views

Field Analysis in Stats

The document is the table of contents for the book 'Discovering Statistics Using IBM SPSS Statistics, 4th Edition' by Andy Field. It outlines various chapters covering topics such as the research process, statistical models, data analysis, and the use of SPSS software for statistical analysis. Each chapter includes sections on theory, practical applications, and further reading.

Uploaded by

nonyethegirl
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 15

DISCOVERING

STATISTICS USING
BM SPSS STATISTICS
4TH EDITION

ANDY FIELD

SSAGE
Los Angeles | London | New Delhi
Singapore | Washington DC
CONTENTS

Preface

How to use this book

Acknowledgements

Dedication

Symbols used in this book

Some maths revision

1 Why is my evil lecturer forcing me to learn statistics?


1.1. What will this chapter tell me? ®
1.2. What the hell am I doing here? I don't belong here ©
1.2.1. The research process ©
1.3. Initial observation: finding something that needs explaining ©
1 4. Generating theories and testing them ©
1.5. Collect data to test your theory ©
1.5.1. Variables©
1.5.2. Measurement error ©
1.5.3. Validity and reliability ©
1.5.4 Correlational research methods ©
1 5.5. Experimental research methods ©
1.5.6. Randomization ©
1.6. Analysing data ©
1.6.1. Frequency distributions ©
1.6.2. The centre of a distribution ©
1.6.3. The dispersion in a distribution ©
1.6.4 Using a frequency distribution to go beyond the data ©
1.6.5. Fitting statistical models to the data ©
1.7 Reporting data ©
1.7.1. Dissemination of research ©
1.7.2. Knowing how to report data ©
1.7.3. Some initial guiding principles ©
1.8. Brian's attempt to woo Jane ©
1.9. What next? ©
1.10. Key terms that I've discovered
1.11. Smart Alex's tasks
1.12. Further reading
DISCOVERING STATISTICS USING IBM SPSS STATISTICS

Everything you never wanted to know about statistics 40


2.1. What will this chapter tell me? © 40
2.2. Building statistical models © 41
2.3. Populations and samples © 42
2.4. Statistical models © 44
2.4.1 The mean as a statistical model © 46
2 4.2. Assessing the fit of a model: sums of squares and variance revisited © 46
2.4.3 Estimating parameters © 50
2.5. Going beyond the data © 51
2.5.1. The standard error © 52
2.5.2. Confidence intervals © 54
2.6. Using statistical models to test research questions © 60
2.61 Null hypothesis significance testing © 60
2.6.2. Problems with NHST © 74
2.7. Modem approaches to theory testing © 78
2.7.1. Effect sizes© 79
2.7.2. Meta-analysis © 83
2.8. Reporting statistical models © 84
2.9. Brian's attempt to woo Jane © 85
2.10. What next?© 86
2.11. Key terms that I've discovered 87
2.12. Smart Alex's tasks 87
2.13. Further reading 88

The IBM SPSS Statistics environment 89


3.1. What will this chapter tell me? © 89
3.2. Versions of IBM SPSS Statistics © 90
3.3. Windows versus MacOS © 90
3.4. Getting started © 90
3.5. The data editor © 91
3.5.1 Entering data into the data editor © 98
3.5.2. The variable view © 99
3.5.3 Missing values © 107
3.6. Importing data © 109
3.7. The SPSS viewer © 109
3.8. Exporting SPSS output © 113
3.9. The syntax editor ® 113
3.10. Saving files © 115
3.11 Retrieving a file © 115
3.12 Brian's attempt to woo Jane © 116
3.13. What next?© 117
3.14 Key terms that I've discovered 117
3.15. Smart Alex's tasks 117
3.16 Further reading 120

Exploring data with graphs 121


4.1. What will this chapter tell me? © 121
4.2. The art of presenting data © 122
4.2.1. What makes a good graph?© 122
4.2.2 Lies, damned lies, and ... erm ... graphs © 123
CONTENTS vii

4.3. The SPSS chart builder © 125


4.4. Histograms© 127
4.5. Boxplots (box-whisker diagrams) © 131
4.6. Graphing means: bar charts and error bars © 135
4.6.1. Simple bar charts for independent means © 136
4 6.2. Clustered bar charts for independent means © 137
4.6.3. Simple bar charts for related means © 140
4.6.4. Clustered bar charts for related means © 143
4.6.5. Clustered bar charts for'mixed'designs © 145
4.7. Line charts© 148
4.8. Graphing relationships: the scatterplot © 148
4.8.1. Simple scatterplot © 149
4.8.2. Grouped scatterplot © 151
4.8.3. Simple and grouped 3-D scatterplots © 153
4.8.4. Matrix scatterplot © 154
4.8.5. Simple dot plot or density plot © 157
4.8.6. Drop-line graph © 157
4.9. Editing graphs © 158
4.10. Brian's attempt to woo Jane © 161
4.11. What next?© 161
4.12. Key terms that I've discovered 161
4.13. Smart Alex's tasks 162
4.14. Further reading 162

5 The beast of bias 163


5.1. What will this chapter tell me? © 163
5.2. What is bias?© 164
5.2.1. Assumptions© 165
5.2.2. Outliers© 165
5.2.3. Additivity and linearity © 167
5.2.4 Normally distributed something or other © 168
5.2.5. Homoscedasticity/homogeneity of variance © 172
5.2.6. Independence © 176
5.3 Spotting bias © 176
5.3.1. Spotting outliers © 176
5.3.2. Spotting normality © 179
5.3.3. Spotting linearity and heteroscedasticity/heterogeneity
of variance© 192
5.4. Reducing bias © 196
5.4.1. Trimming the data © 196
5.4.2. Winsorizing© 198
5.4.3. Robust methods ® 198
5.4.4. Transforming data © 201
5.5. Brian's attempt to woo Jane © 210
5.6. What next?© 210
5.7. Key terms that I've discovered 211
5.8. Smart Alex's tasks 211
5.9. Further reading 212

6 Non-parametric models 213


6.1. What will this chapter tell me? © 213
6.2. When to use non-parametric tests © 214
viii DISCOVERING STATISTICS USING IBM SPSS STATISTICS

6.3. General procedure of non-parametric tests in SPSS © 215


6.4. Comparing two independent conditions: the Wilcoxon rank-sum test and
Mann-Whitney test © 217
6.4.1. Theory© 219
6 4 2. Inputting data and provisional analysis © 221
6.4.3. The Mann-Whitney test using SPSS © 223
6 4 4 Output from the Mann-Whitney test © 224
6.4.5. Calculating an effect size © 227
6.4.6. Writing the results © 227
6.5. Comparing two related conditions:
the Wilcoxon signed-rank test © 228
6.5.1 Theory of the Wilcoxon signed-rank test © 228
6.5.2. Running the analysis © 230
6.5.3. Output for the ecstasy group © 231
6.5.4. Output for the alcohol group © 233
6.5.5. Calculating an effect size © 234
6.5.6. Writing the results © 234
6.6. Differences between several independent groups: the Kruskal-Wallis test © 236
6.6.1 Theory of the Kruskal-Wallis test © 236
6.6.2. Follow-up analysis © 238
6.6.3. Inputting data and provisional analysis © 239
6.6 4. Doing the Kruskal-Wallis test in SPSS © 241
6.6.5. Output from the Kruskal-Wallis test © 242
6.6.6. Testing for trends: the Jonckheere-Terpstra test © 246
6.6.7. Calculating an effect size © 248
6.6.8. Writing and interpreting the results © 249
6.7. Differences between several related groups: Friedman's ANOVA © 249
6.7.1. Theory of Friedman's ANOVA © 251
6.7.2. Inputting data and provisional analysis © 252
6.7.3. Doing Friedman's ANOVA in SPSS © 253
6.7.4. Output from Friedman's ANOVA © 254
6.7.5. Following-up Friedman's ANOVA © 256
6.7.6. Calculating an effect size © 256
6.7.7. Writing and interpreting the results © 257
6.8. Brian's attempt to woo Jane © 258
6.9. What next? © 259
6.10. Key terms that I've discovered 259
6.11. Smart Alex's tasks 259
6.12. Further reading 261

7 Correlation 262
7.1. What will this chapter tell me? © 262
7.2. Modelling relationships © 263
7.2.1. A detour into the murky world of covariance © 264
7.2.2. Standardization and the correlation coefficient © 266
7.2.3. The significance of the correlation coefficient (D 268
7.2 4 Confidence intervals for r © 269
7.2.5. A word of warning about interpretation: causality © 270
7.3 Data entry for correlation analysis using SPSS © 270
7.4. Bivariate correlation © 271
7 4.1 General procedure for running correlations in SPSS © 272
7.4.2. Pearson's correlation coefficient © 274
7 4.3. Spearman's correlation coefficient © 276
CONTENTS

7.4.4. Kendall's tau (non-parametric) © 278


7.4.5. Biserial and point-biserial correlations ® 279
7.5. Partial correlation © 281
7.5.1 The theory behind part and partial correlation ® 281
7.5.2. Partial correlation in SPSS ® 283
7.5.3. Semi-partial (or part) correlations © 285
7.6. Comparing correlations @ 285
7.6.1. Comparing independent rs ® 285
7.6.2. Comparing dependent rs ® 286
7.7. Calculating the effect size © 287
7.8. How to report correlation coefficients © 288
7.9. Brian's attempt to woo Jane © 290
7.10. What next? © 290
7.11. Key terms that I've discovered 291
7.12. Smart Alex's tasks 291
7.13. Further reading 292

8 Regression 293
8.1. What will this chapter tell me? © 293
8.2. An introduction to regression © 294
8.2.1. The simple linear model © 294
8.2.2. The linear model with several predictors © 296
8.2.3. Estimating the model © 298
8.2.4. Assessing the goodness of fit, sums of squares, R and R2 © 300
8.2.5. Assessing individual predictors © 303
8.3. Bias in regression models? © 304
8.3.1. Is the model biased by unusual cases? © 304
8.3.2. Generalizing the model © 309
8.3.3. Sample size in regression ® 313
8.4. Regression using SPSS: One Predictor © 314
8.4.1. Regression: the general procedure © 315
8 4.2. Running a simple regression using SPSS © 316
8.4.3. Interpreting a simple regression © 318
8.4.4. Using the model © 320
8.5. Multiple regression © 321
8.5.1. Methods of regression © 321
8.5.2. Comparing models © 324
8.5.3. Multicollinearity © 324
8.6. Regression with several predictors using SPSS © 326
8.6.1. Main options© 327
8.6.2. Statistics © 328
8.6.3. Regression plots © 329
8.6.4. Saving regression diagnostics © 331
8.6.5. Further options © 332
8.6.6. Robust regression © 333
8.7. Interpreting multiple regression © 334
8.7.1 Descriptives © 334
8.7.2. Summary of model © 335
8.7.3. Model parameters © 338
8.7.4. Excluded variables © 342
8.7.5. Assessing multicollinearity © 342
8.7.6 Bias in the model: casewise diagnostics © 345
8.7.7. Bias in the model: assumptions © 348
X DISCOVERING STATISTICS USING IBM SPSS STATISTICS

8.8. What if I violate an assumption? Robust regression © 350


8.9. How to report multiple regression © 352
8.10. Brian's attempt to woo Jane © 353
8.11. What next? © 354
8.12. Key terms that I've discovered 354
8.13. Smart Alex's tasks 354
8.14. Further reading 356

9 Comparing two means 357


9.1. What will this chapter tell me? © 357
9.2. Looking at differences © 358
9.2.1. An example: are invisible people mischievous? © 359
9.2.2. Categorical predictors in the linear model © 362
9.3. The Mest © 364
9.3.1. Rationale for the Mest © 364
9.3.2. The independent f-test equation explained © 365
9.3.3. The paired-samples Mest equation explained © 368
9.4. Assumptions of the/-test © 371
9.5. The independent Ntest using SPSS © 371
9.5.1. The general procedure © 371
9.5.2. Exploring data and testing assumptions © 372
9.5.3. Compute the independent Mest © 372
9.5.4. Output from the independent Mest © 373
9.5.5. Calculating the effect size © 376
9.5.6. Reporting the independent Mest © 377
9.6. Paired-samples Mest using SPSS © 378
9.6.1. Entering data © 378
9.6.2. Exploring data and testing assumptions © 378
9.6.3. Computing the paired-samples Mest © 383
9 6.4 Calculating the effect size © 386
9.6.5. Reporting the paired-samples Mest © 387
9.7. Between groups or repeated measures? © 388
9.8. What if I violate the test assumptions? © 388
9.9. Brian's attempt to woo Jane © 389
9.10. What next? © 389
9.11. Key terms that I've discovered 389
9.12. Smart Alex's tasks 390
9.13. Further reading 391

10 Moderation, mediation and more regression 392


10.1. What will this chapter tell me? © 392
10.2. Installing custom dialog boxes in SPSS © 393
10.3. Moderation: interactions in regression @ 395
10.3.1 The conceptual model © 395
10.3.2. The statistical model © 397
10.3.3. Centring variables © 398
10.3.4 Creating interaction variables © 400
10.3.5. Following up an interaction effect © 400
10.3.6. Running the analysis © 401
10.3.7. Output from moderation analysis © 402
10.3.8 Reporting moderation analysis © 407
10.4. Mediation © 408
CONTENTS

10.4.1. The conceptual model © 408


10.4.2. The statistical model © 409
10.4.3. Effect sizes of mediation ® 411
10.4.4. Running the analysis © 413
10.4.5. Output from mediation analysis © 414
10.46 Reporting mediation analysis © 418
10.5. Categorical predictors in regression © 419
10.5.1. Dummy coding® 419
10.5.2. SPSS output for dummy variables ® 422
10.6. Brian's attempt to woo Jane © 426
10.7. What next?© 427
10.8. Key terms that I've discovered 427
10.9. Smart Alex's tasks 427
10.10. Further reading 428

11 Comparing several means: ANOVA (GLM 1) 429


11.1. What will this chapter tell me? © 429
11.2. The theory behind ANOVA © 430
11.2.1 Using a linear model to compare means © 430
11.2.2. Logic of the F-ratio © 434
11.2.3 Total sum of squares (SST) © 436
11.2.4 Model sum of squares (SSM) © 438
11.2.5. Residual sum of squares(SSH) © 439
11.2.6. Mean squares © 440
11.2.7. The F-ratio® 441
11.2.8 Interpreting F © 442
11.3. Assumptions of ANOVA © 442
11.3.1. Homogeneity of variance © 442
11.3.2. Is ANOVA robust? ® 444
11.3.3. What to do when assumptions are violated © 445
11.4. Planned contrasts © 445
11.4.1. Choosing which contrasts to do © 446
11.4.2. Defining contrasts using weights © 449
11.4.3. Non-orthogonal comparisons © 454
11.4.4 Standard contrasts © 456
11.4.5 Polynomial contrasts: trend analysis © 457
11.5. Post hoc procedures © 458
11.5.1 Type I and Type II error rates for post hoc tests © 458
11.5.2 Are post hoc procedures robust? © 459
11.5.3 Summary of post hoc procedures © 459
11.6. Running one-way ANOVA in SPSS © 460
11.6.1 General procedure of one-way ANOVA © 460
11.6.2. Planned comparisons using SPSS © 462
11.6.3. Post hoc tests in SPSS © 463
11.6.4. Options © 464
11.6.5 Bootstrapping © 465
11.7. Output from one-way ANOVA © 466
11.7.1 Output for the main analysis © 466
11.7.2. Output for planned comparisons © 469
11.7.3. Output for post hoc tests © 470
11.8. Calculating the effect size © 472
11.9. Reporting results from one-way independent ANOVA © 474
11.10.Key terms that I've discovered 475
DISCOVERING STATISTICS USING IBM SPSS STATISTICS

11.11 Brian's attempt to woo Jane © 475


11.12.What next? © 476
11.13.Smart Alex's tasks 476
11.14. Further reading 477

12 Analysis of covariance, ANCOVA (GLM 2) 478


12.1. What will this chapter tell me? © 478
12.2. What is ANCOVA? © 479
12.3. Assumptions and issues in ANCOVA © 484
12.3.1. Independence of the covariate and treatment effect ® 484
12.3.2 Homogeneity of regression slopes ® 485
12.3.3. What to do when assumptions are violated © 488
12.4. Conducting ANCOVA in SPSS © 488
12.4.1. General procedure © 488
12.4.2. Inputting data © 488
12.4.3. Testing the independence of the treatment variable and covariate © 488
12.4.4 The main analysis © 490
12.4.5. Contrasts 490
12.4.6. Other options © 491
12.4.7. Bootstrapping and plots © 493
12.5. Interpreting the output from ANCOVA © 493
12.5.1 What happens when the covariate is excluded? © 493
12.5.2. The main analysis © 494
12.5.3. Contrasts © 497
12.5.4. Interpreting the covariate © 497
12.6. Testing the assumption of homogeneity of regression slopes © 499
12.7. Calculating the effect size © 500
12.8. Reporting results © 503
12.9. Brian's attempt to woo Jane © 504
12.10. What next?© 504
12.11. Key terms that I've discovered 505
12.12.Smart Alex's tasks 505
12.13.Further reading 506

13 Factorial ANOVA (GLM 3) 507


13.1. What will this chapter tell me? © 507
13.2. Theory of factorial ANOVA (independent designs) © 508
13.2.1 Factorial designs © 508
13.2.2. Guess what? Factorial ANOVA is a linear model © 509
13.2.3. Two-way ANOVA: behind the scenes © 514
13.2.4 Total sums of squares (SST) © 515
13.2.5. Model sum of squares, SSM © 516
13.2.6. The residual sum of squares, SSR © 519
13.2.7. TheF-ratios© 519
13.3. Assumptions of factorial ANOVA © 520
13.4. Factorial ANOVA using SPSS © 520
13.4.1 General procedure for factorial ANOVA © 520
13.4.2. Entering the data and accessing the main dialog box © 521
13.4.3 Graphing interactions © 522
13.4.4. Contrasts© 523
13 4.5. Post hoc tests © 524
13.4.6. Bootstrapping and other options © 524
CONTENTS

13.5. Output from factorial ANOVA © 526


13.5.1. Levene's test © 526
13.5.2. The main ANOVA table © 526
13.5.3. Contrasts© 529
13.5.4. Simple effects analysis ® 530
13.5.5. Post hoc analysis © 532
13.6. Interpreting interaction graphs © 533
13.7. Calculating effect sizes © 537
13.8. Reporting the results of two-way ANOVA © 539
13.9. Brian's attempt to woo Jane © 540
13.10.What next? © 541
13.11.Key terms that I've discovered 541
13.12.Smart Alex's tasks 541
13.13.Further reading 542

14 Repeated-measures designs (GLM 4) 543


14.1. What will this chapter tell me? © 543
14.2. Introduction to repeated-measures designs © 544
14.2.1. The assumption of sphericity © 545
14.2.2. How is sphericity measured? © 545
14.2.3 Assessing the severity of departures from sphericity © 546
14.2.4. What is the effect of violating the assumption of sphericity? © 546
14.2.5. What do you do if you violate sphericity? © 548
14.3. Theory of one-way repeated-measures ANOVA © 548
14.3.1. The total sum of squares, SST © 551
14.3.2. The within-participant sum of squares, SSW © 551
14.3.3. The model sum of squares, SSM © 552
14.3.4 The residual sum of squares, SSR © 553
14.3.5. The mean squares © 553
14.3.6. The F-ratio© 554
14,3.7 The between-participants sum of squares © 554
14 4. Assumptions in repeated-measures ANOVA © 555
14.5. One-way repeated-measures ANOVA using SPSS © 555
14.5.1 Repeated-measures ANOVA: the general procedure 2) 555
14.5.2. The main analysis © 555
14.5.3 Defining contrasts for repeated measures © 557
14.5.4 Post hoc tests and additional options © 558
14.6. Output for one-way repeated-measures ANOVA © 559
14.6.1 Descriptives and other diagnostics © 559
14.6.2 Assessing and correcting for sphericity: Mauchly's test © 560
14.6.3. The main ANOVA © 560
14.6.4 Contrasts® 563
14.6.5. Post hoc tests © 565
14.7. Effect sizes for repeated-measures ANOVA © 566
14.8. Reporting one-way repeated-measures ANOVA © 568
14.9. Factorial repeated-measures designs © 568
14.9.1 The main analysis © 570
14.9.2. Contrasts© 573
14.9.3. Simple effects analysis © 573
14.9.4 Graphing interactions © 574
14.9.5 Other options © 574
14.10. Output for factorial repeated-measures ANOVA © 576
DISCOVERING STATISTICS USING IBM SPSS STATISTICS

14.10.1. Descriptives and main analysis © 576


14.10.2. Contrasts for repeated-measures variables © 581
14.11 Effect sizes for factorial repeated-measures ANOVA ® 586
14.12 Reporting the results from factorial repeated-measures ANOVA © 587
14.13 Brian's attempt to woo Jane © 588
14.14.What next?© 589
14.15.Key terms that I've discovered 589
14.16.Smart Alex's tasks 589
14.17.Further reading 590

Mixed design ANOVA (GLM 5) 591


15.1 What will this chapter tell me? © 591
15.2. Mixed designs © 592
15.3. Assumptions in mixed designs © 593
15.4. What do men and women look for in a partner? © 593
15.5. Mixed ANOVA in SPSS © 594
15.5.1 Mixed ANOVA: the general procedure © 594
15.5.2. Entering data © 594
15.5.3. The main analysis © 596
15.5.4. Other options © 598
15.6. Output for mixed factorial ANOVA © 600
15.6.1. The main effect of gender © 602
15.6.2. The main effect of looks © 603
15.6.3. The main effect of charisma © 605
15.6.4. The interaction between gender and looks © 606
15.6.5. The interaction between gender and charisma © 607
15.6.6. The interaction between attractiveness and charisma © 608
15.6.7 The interaction between looks, charisma and gender ® 611
15.6.8. Conclusions® 614
15.7. Calculating effect sizes © 615
15.8. Reporting the results of mixed ANOVA © 617
15.9. Brian's attempt to woo Jane © 620
15.10What next?© 621
15.11 Key terms that I've discovered 621
15.12.Smart Alex's tasks 621
15.13 Further reading 622

Multivariate analysis of variance (MAN0VA) 623


16 1. What will this chapter tell me? © 623
16.2. When to use MANOVA © 624
16.3. Introduction 624
16.3.1 Similarities to and differences from ANOVA © 624
16.3.2 Choosing outcomes © 625
16.3.3. The example for this chapter © 626
16.4. Theory of MANOVA ® 626
16.4.1. Introduction to matrices © 626
16 4 2. Some important matrices and their functions © 628
16.4.3. Calculating MANOVA by hand: a worked example © 629
16.4 4. Principle of the MANOVA test statistic 0 637
16.5. Practical issues when conducting MANOVA © 642
16.5.1. Assumptions and how to check them © 642
CONTENTS xv

16.5.2 What to do when assumptions are violated ® 643


16.5.3. Choosing a test statistic ® 643
16.5.4. Follow-up analysis ® 644
16.6. MANOVA using SPSS © 644
16.6.1 General procedure of one-way ANOVA © 644
16.6.2. The main analysis © 645
16.6.3. Multiple comparisons in MANOVA © 646
16.6.4. Additional options ® 646
16.7. Output from MANOVA © 647
16.7.1 Preliminary analysis and testing assumptions ® 647
16.7.2. MANOVA test statistics ® 648
16.7.3. Univariate test statistics © 649
16.7.4. SSCP matrices © 650
16.7.5. Contrasts® 652
16.8. Reporting results from MANOVA © 652
16.9. Following up MANOVA with discriminant analysis © 654
16.10.0utput from the discriminant analysis 0 656
16.11 Reporting results from discriminant analysis © 660
16.12.The final interpretation © 660
16.13.Brian's attempt to woo Jane © 662
16.14.What next?© 663
16.15.Key terms that I've discovered 663
16.16.Smart Alex's tasks 664
16.17.Further reading 664

17 Exploratory factor analysis 665


17.1. What will this chapter tell me? © 665
17.2. When to use factor analysis © 666
17.3. Factors and components © 667
17.3.1. Graphical representation © 668
17.3.2. Mathematical representation © 669
17.3.3. Factor scores © 671
17.4. Discovering factors © 674
17.4.1. Choosing a method © 674
17.4.2. Communality© 675
17.4.3. Factor analysis or PCA? © 675
17.4.4. Theory behind PCA ® 676
17.4.5 Factor extraction: eigenvalues and the scree plot © 677
17.4.6. Improving interpretation: factor rotation © 678
17.5. Research example © 682
17.5.1. General procedure © 682
17.5.2. Before you begin © 683
17.6. Running the analysis © 686
; 17.6.1 Factor extraction in SPSS © 688
17.6.2. Rotation© 689
17.6.3. Scores® 691
17.6.4. Options© 691
17.7. Interpreting output from SPSS © 692
17.7.1. Preliminary analysis © 693
17.7.2. Factor extraction © 696
17.7.3. Factor rotation © 701
17.7.4. Factor scores © 704
17.7.5. Summary® 705
xvi DISCOVERING STATISTICS USING IBM SPSS STATISTICS

17.8. How to report factor analysis ® 706


17.9. Reliability analysis © 706
17.9.1. Measures of reliability © 706
17 9.2. Interpreting Cronbach's a (some cautionary tales) © 709
17.9.3. Reliability analysis in SPSS © 710
17.9.4. Reliability analysis output © 712
17.10.How to report reliability analysis © 716
17.11 Brian's attempt to woo Jane © 716
17.12.What next? © 717
17.13.Key terms that I've discovered 717
17.14.Smart Alex's tasks 717
17.15.Further reading 719

18 Categorical data 720


18.1. What will this chapter tell me? © 720
18.2. Analysing categorical data © 721
18.3. Theory of analysing categorical data © 721
18.3.1. Pearson's chi-square test © 721
18.3.2. Fisher's exact test © 723
18.3.3. The likelihood ratio © 724
18.3.4. Yates's correction © 724
18.3.5. Other measures of association © 725
18.3.6 Several categorical variables: loglinear analysis © 725
18.4. Assumptions when analysing categorical data © 735
18.4.1. Independence© 735
18.4.2. Expected frequencies © 735
18.4.3. More doom and gloom © 736
18.5. Doing chi-square in SPSS © 736
18 5.1 General procedure for analysing categorical outcomes © 736
18 5.2. Entering data © 736
18.5.3. Running the analysis © 738
18.5.4. Output for the chi-square test © 740
18.5.5. Breaking down a significant chi-square test with standardized residuals © 743
18.5.6. Calculating an effect size © 744
18.5.7. Reporting the results of chi-square © 746
18.6. Loglinear analysis using SPSS © 746
18.6.1. Initial considerations © 746
18.6.2. Running loglinear analysis © 748
18.6.3. Output from loglinear analysis © 750
18.6 4 Following up loglinear analysis © 753
18.7. Effect sizes in loglinear analysis © 755
18.8. Reporting the results of loglinear analysis © 756
18.9. Brian's attempt to woo Jane © 757
18.10.What next?© 757
18.11 Key terms that I've discovered 758
18.12.Smart Alex's tasks 758
18.13.Further reading 759

19 Logistic regression 760


19.1. What will this chapter tell me? © 760
19.2. Background to logistic regression © 761
CONTENTS xvii

19.3. What are the principles behind logistic regression? ® 762


19.3.1 Assessing the model: the log-likelihood statistic ® 763
19.3.2 Assessing the model: the deviance statistic ® 763
19.3.3 Assessing the model: R and R2 ® 764
19.3.4 Assessing the contribution of predictors: the Wald statistic ® 766
19.3.5. The odds ratio: exp(S) ® 766
19.3.6. Model building and parsimony © 767
19.4. Sources of bias and common problems © 768
19.4.1. Assumptions© 768
19.4.2 Incomplete information from the predictors © 769
19.4.3. Complete separation © 770
19.4.4. Overdispersion © 772
19.5. Binary logistic regression: an example that will make you feel eel © 773
19.5.1. Building a model © 774
19.5.2. Logistic regression: the general procedure © 775
19.5.3. Data entry© 775
19.5.4. Building the models in SPSS © 775
19.5.5. Method of regression © 776
19.5.6. Categorical predictors © 776
19.5.7. Comparing the models © 778
19.5.8. Rerunning the model © 780
19.5.9. Obtaining residuals © 781
19.5.10. Further options © 781
19.5.11 Bootstrapping© 782
19.6. Interpreting logistic regression © 783
19.6.1. Block 0© 783
19.6.2. Model summary © 783
19.6.3. Listing predicted probabilities © 789
19.6.4 Interpreting residuals ® 789
19.6.5. Calculating the effect size © 792
19.7. How to report logistic regression © 792
19.8. Testing assumptions: another example © 792
19.8.1. Testing for linearity of the logit © 794
19.8.2. Testing for multicollinearity ® 794
19.9. Predicting several categories: multinomial logistic regression ® 797
19.9.1 Running multinomial logistic regression in SPSS ® 799
19.9.2. Statistics ® 802
19.9.3. Other options ® 803
19.9.4 Interpreting the multinomial logistic regression output ® 804
19.9.5. Reporting the results © 811
19.10.Brian's attempt to woo Jane © 811
19.11.What next?© 811
19.12.Key terms that I've discovered 812
19.13.Smart Alex's tasks 812
19.14.Further reading 813

20 Multilevel linear models 814


20.1. What will this chapter tell me? © 814
20.2. Hierarchical data © 815
20.2.1. The intraclass correlation © 816
20.2.2. Benefits of multilevel models © 818
xviii DISCOVERING STATISTICS USING IBM SPSS STATISTICS

20.3 Theory of multilevel linear models ® 819


20.3.1. An example© 819
20.3.2. Fixed and random coefficients © 820
20.4 The multilevel model © 823
20.4.1 Assessing the fit and comparing multilevel models © 825
20.4.2. Types of covariance structures © 826
20.5 Some practical issues ® 827
20.5.1. Assumptions® 827
20.5.2. Robust multilevel models © 828
20.5.3. Sample size and power © 829
20.5.4. Centring predictors © 829
20.6 Multilevel modelling using SPSS © 830
20.6.1. Entering the data © 831
20.6.2 Ignoring the data structure: ANOVA © 831
20.6.3 Ignoring the data structure: ANCOVA © 836
20.6.4 Factoring in the data structure: random intercepts ® 837
20.6.5 Factoring in the data structure: random intercepts and slopes © 841
20.6.6 Adding an interaction to the model © 845
20.7. Growth models © 849
20.7.1. Growth curves (polynomials) © 850
20.7.2. An example: the honeymoon period © 851
20.7.3. Restructuring the data ® 853
20 7.4 Running a growth model on SPSS © 854
20.7.5. Further analysis © 860
20.8. How to report a multilevel model ® 862
20.9. A message from the octopus of inescapable despair © 863
20.10.Brian's attempt to woo Jane © 864
20.11.What next? ® 864
20.12 Key terms that I've discovered 865
20.13 Smart Alex's tasks 865
20.14 Further reading 866

21 Epilogue: life after discovering statistics 867


21.1. Nice emails 867
21.2. Everybody thinks that I'm a statistician 868
21.3. Craziness on a grand scale 868
21.3.1. Catistics 868
21.3.2. Cult of underlying numerical truths 869
21.3.3. And then it got really weird 869

Glossary 870

Appendix 887

References 899

Index 908

You might also like