Log In
Or create an account ->
Imperial Library
Home
About
News
Upload
Forum
Help
Login/SignUp
Index
Title Page
Copyright Page
Contents
Preface
How to use this book
Acknowledgements
Dedication
Symbols used in this book
Some maths revision
1 Why is my evil lecturer forcing me to learn statistics?
1.1. What will this chapter tell me?
1.2. What the hell am I doing here? I don’t belong here
1.3. Initial observation: finding something that needs explaining
1.4. Generating theories and testing them
1.5. Data collection 1: what to measure
1.5.1. Variables
1.5.2. Measurement error
1.5.3. Validity and reliability
1.6. Data collection 2: how to measure
1.6.1. Correlational research methods
1.6.2. Experimental research methods
1.6.3. Randomization
1.7. Analysing data
1.7.1. Frequency distributions
1.7.2. The centre of a distribution
1.7.3. The dispersion in a distribution
1.7.4. Using a frequency distribution to go beyond the data
1.7.5. Fitting statistical models to the data
What have I discovered about statistics?
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
2 Everything you ever wanted to know about statistics(well, sort of) (well, sort of)
2.1. What will this chapter tell me?
2.2. Building statistical models
2.3. Populations and samples
2.4. Simple statistical models
2.4.1. The mean: a very simple statistical model
2.4.2. Assessing the fit of the mean: sums of squares, variance and standard deviations
2.4.3. Expressing the mean as a model
2.5. Going beyond the data
2.5.1. The standard error
2.5.2. Confidence intervals
2.6. Using statistical models to test research questions
2.6.1. Test statistics
2.6.2. One- and two-tailed tests
2.6.3. Type I and Type II errors
2.6.4. Effect sizes
2.6.5. Statistical power
What have I discovered about statistics?
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
3 The R environment
3.1. What will this chapter tell me?
3.2. Before you start
3.2.1. The R-chitecture
3.2.2. Pros and cons of R
3.2.3. Downloading and installing R
3.2.4. Versions of R
3.3. Getting started
3.3.1. The main windows in R
3.3.2. Menus in R
3.4. Using R
3.4.1. Commands, objects and functions
3.4.2. Using scripts
3.4.3. The R workspace
3.4.4. Setting a working directory
3.4.5. Installing packages
3.4.6. Getting help
3.5. Getting data into R
3.5.1. Creating variables
3.5.2. Creating dataframes
3.5.3. Calculating new variables from exisiting ones
3.5.4. Organizing your data
3.5.5. Missing values
3.6. Entering data with R Commander
3.6.1. Creating variables and entering data with R Commander
3.6.2. Creating coding variables with R Commander
3.7. Using other software to enter and edit data
3.7.1. Importing data
3.7.2. Importing SPSS data files directly
3.7.3. Importing data with R Commander
3.7.4. Things that can go wrong
3.8. Saving data
3.9. Manipulating data
3.9.1. Selecting parts of a dataframe
3.9.2. Selecting data with the subset() function
3.9.3. Dataframes and matrices
3.9.4. Reshaping data
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
4 Exploring data with graphs
4.1. What will this chapter tell me?
4.2. The art of presenting data
4.2.1. Why do we need graphs
4.2.2. What makes a good graph?
4.2.3. Lies, damned lies, and … erm … graphs
4.3. Packages used in this chapter
4.4. Introducing ggplot2
4.4.1. The anatomy of a plot
4.3.2. Geometric objects (geoms)
4.4.3. Aesthetics
4.4.4. The anatomy of the ggplot() function
4.4.5. Stats and geoms
4.4.6. Avoiding overplotting
4.4.7. Saving graphs
4.4.8. Putting it all together: a quick tutorial
4.5. Graphing relationships: the scatterplot
4.5.1. Simple scatterplot
4.5.2. Adding a funky line
4.5.3. Grouped scatterplot
4.6. Histograms: a good way to spot obvious problems
4.7. Boxplots (box–whisker diagrams)
4.8. Density plots
4.9. Graphing means
4.9.1. Bar charts and error bars
4.9.2. Line graphs
4.10. Themes and options
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
5 Exploring assumptions
5.1. What will this chapter tell me?
5.2. What are assumptions?
5.3. Assumptions of parametric data
5.4. Packages used in this chapter
5.5. The assumption of normality
5.5.1. Oh no, it’s that pesky frequency distribution again: checking normality visually
5.5.2. Quantifying normality with numbers
5.5.3. Exploring groups of data
5.6. Testing whether a distribution is normal
5.6.1. Doing the Shapiro–Wilk test in R
5.6.2. Reporting the Shapiro–Wilk test
5.7. Testing for homogeneity of variance
5.7.1. Levene’s test
5.7.2. Reporting Levene’s test
5.7.3. Hartley’s Fmax: the variance ratio
5.8. Correcting problems in the data
5.8.1. Dealing with outliers
5.8.2. Dealing with non-normality and unequal variances
5.8.3. Transforming the data using R
5.8.4. When it all goes horribly wrong
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
6 Correlation
6.1. What will this chapter tell me?
6.2. Looking at relationships
6.3. How do we measure relationships?
6.3.1. A detour into the murky world of covariance
6.3.2. Standardization and the correlation coefficient
6.3.3. The significance of the correlation coefficient
6.3.4. Confidence intervals for r
6.3.5. A word of warning about interpretation: causality
6.4. Data entry for correlation analysis
6.5. Bivariate correlation
6.5.1. Packages for correlation analysis in R
6.5.2. General procedure for correlations using R Commander
6.5.3. General procedure for correlations using R
6.5.4. Pearson’s correlation coefficient
6.5.5. Spearman’s correlation coefficient
6.5.6. Kendall’s tau (non-parametric)
6.5.7. Bootstrapping correlations
6.5.8. Biserial and point-biserial correlations
6.6. Partial correlation
6.6.1. The theory behind part and partial correlation
6.6.2. Partial correlation using R
6.6.3 Semi-partial (or part) correlations
6.7. Comparing correlations
6.7.1. Comparing independent rs
6.7.2. Comparing dependent rs
6.8. Calculating the effect size
6.9. How to report correlation coefficents
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
7 Regression
7.1. What will this chapter tell me?
7.2. An introduction to regression
7.2.1. Some important information about straight lines
7.2.2. The method of least squares
7.2.3. Assessing the goodness of fit: sums of squares, R and R2
7.2.4. Assessing individual predictors
7.3. Packages used in this chapter
7.4. General procedure for regression in R
7.4.1. Doing simple regression using R Commander
7.4.2. Regression in R
7.5. Interpreting a simple regression
7.5.1. Overall fit of the object model
7.5.2. Model parameters
7.5.3. Using the model
7.6. Multiple regression: the basics
7.6.1. An example of a multiple regression model
7.6.2. Sums of squares, R and R2
7.6.3. Parsimony-adjusted measures of fit
7.6.4. Methods of regression
7.7. How accurate is my regression model?
7.7.1. Assessing the regression model I: diagnostics
7.7.2. Assessing the regression model II: generalization
7.8. How to do multiple regression using R Commander and R
7.8.1. Some things to think about before the analysis
7.8.2. Multiple regression: running the basic model
7.8.3. Interpreting the basic multiple regression
7.8.4. Comparing models
7.9. Testing the accuracy of your regression model
7.9.1. Diagnostic tests using R Commander
7.9.2. Outliers and influential cases
7.9.3. Assessing the assumption of independence
7.9.4. Assessing the assumption of no multicollinearity
7.9.5. Checking assumptions about the residuals
7.9.6. What if I violate an assumption?
7.10. Robust regression: bootstrapping
7.11. How to report multiple regression
7.12. Categorical predictors and multiple regression
7.12.1. Dummy coding
7.12.2. Regression with dummy variables
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
8 Logistic regression
8.1. What will this chapter tell me?
8.2. Background to logistic regression
8.3. What are the principles behind logistic regression?
8.3.1. Assessing the model: the log-likelihood statistic
8.3.2. Assessing the model: the deviance statistic
8.3.3. Assessing the model: R and R2
8.3.4. Assessing the model: information criteria
8.3.5. Assessing the contribution of predictors: the z-statistic
8.3.6. The odds ratio
8.3.7. Methods of logistic regression
8.4. Assumptions and things that can go wrong
8.4.1. Assumptions
8.4.2. Incomplete information from the predictors
8.4.3. Complete separation
8.5. Packages used in this chapter
8.6. Binary logistic regression: an example that will make you feel eel
8.6.1. Preparing the data
8.6.2. The main logistic regression analysis
8.6.3. Basic logistic regression analysis using R
8.6.4. Interpreting a basic logistic regression
8.6.5. Model 1: Intervention only
8.6.6. Model 2: Intervention and Duration as predictors
8.6.7. Casewise diagnostics in logistic regression
8.6.8. Calculating the effect size
8.7. How to report logistic regression
8.8. Testing assumptions: another example
8.8.1. Testing for multicollinearity
8.8.2. Testing for linearity of the logit
8.9. Predicting several categories: multinomial logistic regression
8.9.1. Running multinomial logistic regression in R
8.9.2. Interpreting the multinomial logistic regression output
8.9.3. Reporting the results
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
9 Comparing two means
9.1. What will this chapter tell me?
9.2. Packages used in this chapter
9.3. Looking at differences
9.3.1. A problem with error bar graphs of repeated-measures designs
9.3.2. Step 1: calculate the mean for each participant
9.3.3. Step 2: calculate the grand mean
9.3.4. Step 3: calculate the adjustment factor
9.3.5. Step 4: create adjusted values for each variable
9.4. The t-test
9.4.1. Rationale for the t-test
9.4.2. The t-test as a general linear model
9.4.3. Assumptions of the t-test
9.5. The independent t-test
9.5.1. The independent t-test equation explained
9.5.2. Doing the independent t-test
9.6. The dependent t-test
9.6.1. Sampling distributions and the standard error
9.6.2. The dependent t-test equation explained
9.6.3. Dependent t-tests using R
9.7. Between groups or repeated measures?
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
10 Comparing several means: ANOVA (GLM 1)
10.1. What will this chapter tell me?
10.2. The theory behind ANOVA
10.2.1 Inflated error rates
10.2.2. Interpreting F
10.2.3. ANOVA as regression
10.2.4. Logic of the F-ratio
10.2.5. Total sum of squares (SST)
10.2.6. Model sum of squares (SSM)
10.2.7. Residual sum of squares (SSR)
10.2.8. Mean squares
10.2.9. The F-ratio
10.3. Assumptions of ANOVA
10.3.1. Homogeneity of variance
10.3.2. Is ANOVA robust?
10.4. Planned contrasts
10.4.1. Choosing which contrasts to do
10.4.2. Defining contrasts using weights
10.4.3. Non-orthogonal comparisons
10.4.4. Standard contrasts
10.4.5. Polynomial contrasts: trend analysis
10.5. Post hoc procedures
10.5.1. Post hoc procedures and Type I (α) and Type II error rates
10.5.2. Post hoc procedures and violations of test assumptions
10.5.3. Summary of post hoc procedures
10.6. One-way ANOVA using R
10.6.1. Packages for one-way ANOVA in R
10.6.2. General procedure for one-way ANOVA
10.6.3. Entering data
10.6.4. One-way ANOVA using R Commander
10.6.5. Exploring the data
10.6.6. The main analysis
10.6.7. Planned contrasts using R
10.6.8. Post hoc tests using R
10.7. Calculating the effect size
10.8. Reporting results from one-way independent ANOVA
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
11 Analysis of covariance, ANCOVA (GLM 2)
11.1. What will this chapter tell me?
11.2. What is ANCOVA?
11.3. Assumptions and issues in ANCOVA
11.3.1. Independence of the covariate and treatment effect
11.3.2. Homogeneity of regression slopes
11.4. ANCOVA using R
11.4.1. Packages for ANCOVA in R
11.4.2. General procedure for ANCOVA
11.4.3. Entering data
11.4.4. ANCOVA using R Commander
11.4.5. Exploring the data
11.4.6. Are the predictor variable and covariate independent?
11.4.7. Fitting an ANCOVA model
11.4.8. Interpreting the main ANCOVA model
11.4.9. Planned contrasts in ANCOVA
11.4.10. Interpreting the covariate
11.4.11. Post hoc tests in ANCOVA
11.4.12. Plots in ANCOVA
11.4.13. Some final remarks
11.4.14. Testing for homogeneity of regression slopes
11.5. Robust ANCOVA
11.6. Calculating the effect size
11.7. Reporting results
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
12 Factorial ANOVA (GLM 3)
12.1. What will this chapter tell me?
12.2. Theory of factorial ANOVA (independent design)
12.2.1. Factorial designs
12.3. Factorial ANOVA as regression
12.3.1. An example with two independent variables
12.3.2. Extending the regression model
12.4. Two-way ANOVA: behind the scenes
12.4.1. Total sums of squares (SST)
12.4.2. The model sum of squares (SSM)
12.4.3. The residual sum of squares (SSR)
12.4.4. The F-ratios
12.5. Factorial ANOVA using R
12.5.1. Packages for factorial ANOVA in R
12.5.2. General procedure for factorial ANOVA
12.5.3. Factorial ANOVA using R Commander
12.5.4. Entering the data
12.5.5. Exploring the data
12.5.6. Choosing contrasts
12.5.7. Fitting a factorial ANOVA model
12.5.8. Interpreting factorial ANOVA
12.5.9. Interpreting contrasts
12.5.10. Simple effects analysis
12.5.11. Post hoc analysis
12.5.12. Overall conclusions
12.5.13. Plots in factorial ANOVA
12.6. Interpreting interaction graphs
12.7. Robust factorial ANOVA
12.8. Calculating effect sizes
12.9. Reporting the results of two-way ANOVA
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
13 Repeated-measures designs (GLM 4)
13.1. What will this chapter tell me?
13.2. Introduction to repeated-measures designs
13.2.1. The assumption of sphericity
13.2.2. How is sphericity measured?
13.2.3. Assessing the severity of departures from sphericity
13.2.4. What is the effect of violating the assumption of sphericity?
13.2.5. What do you do if you violate sphericity?
13.3. Theory of one-way repeated-measures ANOVA
13.3.1. The total sum of squares (SST)
13.3.2. The within-participant sum of squares (SSW)
13.3.3. The model sum of squares (SSM)
13.3.4. The residual sum of squares (SSR)
13.3.5. The mean squares
13.3.6. The F-ratio
13.3.7. The between-participant sum of squares
13.4. One-way repeated-measures designs using R
13.4.1. Packages for repeated measures designs in R
13.4.2. General procedure for repeated-measures designs
13.4.3. Repeated-measures ANOVA using R Commander
13.4.4. Entering the data
13.4.5. Exploring the data
13.4.6. Choosing contrasts
13.4.7. Analysing repeated measures: two ways to skin a .dat
13.4.8. Robust one-way repeated-measures ANOVA
13.5. Effect sizes for repeated-measures designs
13.6. Reporting one-way repeated-measures designs
13.7. Factorial repeated-measures designs
13.7.1. Entering the data
13.7.2. Exploring the data
13.7.3. Setting contrasts
13.7.4. Factorial repeated-measures ANOVA
13.7.5. Factorial repeated-measures designs as a GLM
13.7.6. Robust factorial repeated-measures ANOVA
13.8. Effect sizes for factorial repeated-measures designs
13.9. Reporting the results from factorial repeated-measures designs
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
14 Mixed designs (GLM 5)
14.1. What will this chapter tell me?
14.2. Mixed designs
14.3. What do men and women look for in a partner?
14.4. Entering and exploring your data
14.4.1. Packages for mixed designs in R
14.4.2. General procedure for mixed designs
14.4.3. Entering the data
14.4.4. Exploring the data
14.5. Mixed ANOVA
14.6. Mixed designs as a GLM
14.6.1. Setting contrasts
14.6.2. Building the model
14.6.3. The main effect of gender
14.6.4. The main effect of looks
14.6.5. The main effect of personality
14.6.6. The interaction between gender and looks
14.6.7. The interaction between gender and personality
14.6.8. The interaction between looks and personality
14.6.9. The interaction between looks, personality and gender
14.6.10. Conclusions
14.7. Calculating effect sizes
14.8. Reporting the results of mixed ANOVA
14.9. Robust analysis for mixed designs
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
15 Non-parametric tests
15.1. What will this chapter tell me?
15.2. When to use non-parametric tests
15.3. Packages used in this chapter
15.4. Comparing two independent conditions: the Wilcoxon rank-sum test
15.4.1. Theory of the Wilcoxon rank-sum test
15.4.2. Inputting data and provisional analysis
15.4.3. Running the analysis using R Commander
15.4.4. Running the analysis using R
15.4.5. Output from the Wilcoxon rank-sum test
15.4.6. Calculating an effect size
15.4.7. Writing the results
15.5. Comparing two related conditions: the Wilcoxon signed-rank test
15.5.1. Theory of the Wilcoxon signed-rank test
15.5.2. Running the analysis with R Commander
15.5.3. Running the analysis using R
15.5.4. Wilcoxon signed-rank test output
15.5.5. Calculating an effect size
15.5.6. Writing the results
15.6. Differences between several independent groups: the Kruskal–Wallis test
15.6.1. Theory of the Kruskal–Wallis test
15.6.2. Inputting data and provisional analysis
15.6.3. Doing the Kruskal–Wallis test using R Commander
15.6.4. Doing the Kruskal–Wallis test using R
15.6.5. Output from the Kruskal–Wallis test
15.6.6. Post hoc tests for the Kruskal–Wallis test
15.6.7. Testing for trends: the Jonckheere–Terpstra test
15.6.8. Calculating an effect size
15.6.9. Writing and interpreting the results
15.7. Differences between several related groups: Friedman’s ANOVA
15.7.1. Theory of Friedman’s ANOVA
15.7.2. Inputting data and provisional analysis
15.7.3. Doing Friedman’s ANOVA in R Commander
15.7.4. Friedman’s ANOVA using R
15.7.5. Output from Friedman’s ANOVA
15.7.6. Post hoc tests for Friedman’s ANOVA
15.7.7. Calculating an effect size
15.7.8. Writing and interpreting the results
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
16 Multivariate analysis of variance (MANOVA)
16.1. What will this chapter tell me?
16.2. When to use MANOVA
16.3. Introduction: similarities to and differences from ANOVA
16.3.1. Words of warning
16.3.2. The example for this chapter
16.4. Theory of MANOVA
16.4.1. Introduction to matrices
16.4.2. Some important matrices and their functions
16.4.3. Calculating MANOVA by hand: a worked example
16.4.4. Principle of the MANOVA test statistic
16.5. Practical issues when conducting MANOVA
16.5.1. Assumptions and how to check them
16.5.2. Choosing a test statistic
16.5.3. Follow-up analysis
16.6. MANOVA using R
16.6.1. Packages for factorial ANOVA in R
16.6.2. General procedure for MANOVA
16.6.3. MANOVA using R Commander
16.6.4. Entering the data
16.6.5. Exploring the data
16.6.6. Setting contrasts
16.6.7. The MANOVA model
16.6.8. Follow-up analysis: univariate test statistics
16.6.9. Contrasts
16.7. Robust MANOVA
16.8. Reporting results from MANOVA
16.9. Following up MANOVA with discriminant analysis
16.10. Reporting results from discriminant analysis
16.11. Some final remarks
16.11.1. The final interpretation
16.11.2. Univariate ANOVA or discriminant analysis?
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
17 Exploratory factor analysis
17.1. What will this chapter tell me?
17.2. When to use factor analysis
17.3. Factors
17.3.1. Graphical representation of factors
17.3.2. Mathematical representation of factors
17.3.3. Factor scores
17.3.4. Choosing a method
17.3.5. Communality
17.3.6. Factor analysis vs. principal components analysis
17.3.7. Theory behind principal components analysis
17.3.8. Factor extraction: eigenvalues and the scree plot
17.3.9. Improving interpretation: factor rotation
17.4. Research example
17.4.1. Sample size
17.4.2. Correlations between variables
17.4.3. The distribution of data
17.5. Running the analysis with R Commander
17.6. Running the analysis with R
17.6.1. Packages used in this chapter
17.6.2. Initial preparation and analysis
17.6.3. Factor extraction using R
17.6.4. Rotation
17.6.5. Factor scores
17.6.6. Summary
17.7. How to report factor analysis
17.8. Reliability analysis
17.8.1. Measures of reliability
17.8.2. Interpreting Cronbach’s α (some cautionary tales …)
17.8.3. Reliability analysis with R Commander
17.8.4. Reliability analysis using R
17.8.5. Interpreting the output
17.9. Reporting reliability analysis
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
18 Categorical data
18.1. What will this chapter tell me?
18.2. Packages used in this chapter
18.3. Analysing categorical data
18.4. Theory of analysing categorical data
18.4.1. Pearson’s chi-square test
18.4.2. Fisher’s exact test
18.4.3. The likelihood ratio
18.4.4. Yates’s correction
18.5. Assumptions of the chi-square test
18.6. Doing the chi-square test using R
18.6.1. Entering data: raw scores
18.6.2. Entering data: the contingency table
18.6.3. Running the analysis with R Commander
18.6.4. Running the analysis using R
18.6.5. Output from the CrossTable() function
18.6.6. Breaking down a significant chi-square test with standardized residuals
18.6.7. Calculating an effect size
18.6.8. Reporting the results of chi-square
18.7. Several categorical variables: loglinear analysis
18.7.1. Chi-square as regression
18.7.2. Loglinear analysis
18.8. Assumptions in loglinear analysis
18.9. Loglinear analysis using R
18.9.1. Initial considerations
18.9.2. Loglinear analysis as a chi-square test
18.9.3. Output from loglinear analysis as a chi-square test
18.9.4. Loglinear analysis
18.10. Following up loglinear analysis
18.11. Effect sizes in loglinear analysis
18.12. Reporting the results of loglinear analysis
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
19 Multilevel linear models
19.1. What will this chapter tell me?
19.2. Hierarchical data
19.2.1. The intraclass correlation
19.2.2. Benefits of multilevel models
19.3. Theory of multilevel linear models
19.3.1. An example
19.3.2. Fixed and random coefficients
19.4. The multilevel model
19.4.1. Assessing the fit and comparing multilevel models
19.4.2. Types of covariance structures
19.5. Some practical issues
19.5.1. Assumptions
19.5.2. Sample size and power
19.5.3. Centring variables
19.6. Multilevel modelling in R
19.6.1. Packages for multilevel modelling in R
19.6.2. Entering the data
19.6.3. Picturing the data
19.6.4. Ignoring the data structure: ANOVA
19.6.5. Ignoring the data structure: ANCOVA
19.6.6. Assessing the need for a multilevel model
19.6.7. Adding in fixed effects
19.6.8. Introducing random slopes
19.6.9. Adding an interaction term to the model
19.7. Growth models
19.7.1. Growth curves (polynomials)
19.7.2. An example: the honeymoon period
19.7.3. Restructuring the data
19.7.4. Setting up the basic model
19.7.5. Adding in time as a fixed effect
19.7.6. Introducing random slopes
19.7.7. Modelling the covariance structure
19.7.8. Comparing models
19.7.9. Adding higher-order polynomials
19.7.10. Further analysis
19.8. How to report a multilevel model
What have I discovered about statistics?
R packages used in this chapter
R functions used in this chapter
Key terms that I’ve discovered
Smart Alex’s tasks
Further reading
Interesting real research
Epilogue: life after discovering statistics
Troubleshooting R
Glossary
Appendix
A.1. Table of the standard normal distribution
A.2. Critical values of the t-distribution
A.3. Critical values of the F-distribution
A.4. Critical values of the chi-square distribution
References
Index
Functions in R
Packages in R
← Prev
Back
Next →
← Prev
Back
Next →