Multiple Regression ANOVA
Multiple Regression ANOVA
Multiple Regression ANOVA
Regression allows you to investigate the relationship between variables. But more than that, it
allows you to model the relationship between variables, which enables you to make predictions
about what one variable will do based on the scores of some other variables.
• The variable you want to predict is called the outcome variable (or DV)
• The variables you base your prediction on are called the predictor variables (or IVs)
While simple linear regression only enables you to predict the value of one variable based on
the value of a single predictor variable; multiple regression allows you to use multiple
predictors.
Worked Example
For this tutorial, we will use an example based on a fictional study attempting to model
students’ exam performance.
Imagine you are a psychology research methods tutor interested in predicting how well your
students will do in their exam. You think that revision intensity and enjoyment of the subject
are variables that may allow you to do this.
To investigate this you could measure how many hours of revision your students did in the
weeks preceding their exam and ask them to rate their enjoyment of the material on a scale
from 0-100. You could then see how well they do in their exam, which would allow you to
model how well future students are likely to do based on these predictors. This is what we will
explore in this tutorial.
Multiple regression allows you to include multiple predictors (IVs) into your predictive model,
however this tutorial will concentrate on the simplest type: when you have only two predictors
and a single outcome (DV) variable.
As with ANOVA there are a number of assumptions that must be met for multiple regression to
be reliable, however this tutorial only covers how to run the analysis. If you plan on running a
multiple regression as part of your own research project, make sure you also check out the
assumptions tutorial.
This what the data looks like in SPSS. It can also be found in the SPSS file: ‘Week 6 MR
Data.sav’.
In multiple regression, each participant provides a score for all of the variables. As each row
should contain all of the information provided by one participant, there needs to be a separate
column for each variable.
This opens the Linear Regression dialog box. Here you will see all of the variables recorded in
the data file displayed in the box in the left. To tell SPSS what we want to analyse we need to
move our variables to the correct boxes on the right.
Exam_Score is already selected. As this is out Outcome Variable, move it across to the
Dependent box by CLICKING the arrow to the left of it.
Next, SELECT the two predictor variables (Revision Intensity and Subject Enjoyment) as shown
below. When doing this yourself, remember that if you hold down the Ctrl key so you can
highlight them all in one go.
Add them to the analysis by CLICKING on the blue arrow to the left of the Independent(s) box.
Now we have told SPSS which variables are which, we need to tell it what statistics we want it
to produce.
This tutorial will now take you through the results, box-by-box.
Descriptive Statistics
The next box gives you the correlations between each of the variables. The first row shows the
correlation coefficients (‘r ’), while the second tells you their statistical significance. To
establish which values are associated with which correlations you can find the name of the first
variable at the top of each column, and the name of the correlated variable at the start of the
intersecting row.
In multiple regression, you want the predictor variables to be related to your outcome variable
(otherwise, there is no point in including them in the predictive model). In contrast, you don’t
want your predictors to be too strongly related to one another, as this can make your analysis
unreliable. When predictors correlate at more than r = .8, you have multicollinearity which is a
problem for multiple regression, so you may want to remove one of the variables. You can
learn more about this in the separate tutorials on Assumptions of Multiple Regression. In this
case there are several correlations of around r = .5, suggesting multiple regression is
appropriate.
Variables Entered/Removed
Model Summary
The next box displays information about how the two variables relate to one another. In this
case, the term ‘model’ is used because we are trying to build a model of the relationship
between our variables. The model consists of the predictor variables we are using to try to
predict the outcome variable (Exam Score). In this case, we have two predictor variables in the
model: Revision Intensity and Subject Enjoyment.
• R The value in the R column is a very similar statistic to r, and can be interpreted like
any regular correlation coefficient. But instead of telling you the relationship between
two variables, it tells you the strength of the relationship between the outcome
variable (DV) and all of the predictor variables (or IVs) combined.
In this case R = 0.65, which is a strong relationship. This suggests our model is a
relatively good predictor of the outcome.
• R Square The R Square column contains the value we are most interested in. Usually
written as R2, this value indicates the proportion of variation in the outcome variable
(Exam Score) that can be explained by the model (i.e. by Revision Intensity and Subject
Enjoyment).
You can either report this as R2 = .418, or you can multiply it by 100 to give a
proportion. In this case we could say that 41.8% of the variance in the data can be
explained by the predictor variables.
ANOVA
The next box in the output tells us whether or not our model (which includes Revision Intensity
and Subject Enjoyment) is a significant predictor of the outcome variable. This is tested using
Analysis of Variance. As the significance value is less than p=0.05, we can say that the
regression model significantly predicts Exam Score.
So we know that the model is significant, but how do we write up the numbers? To report your
findings in APA format, you report your results as:
You need to report these statistics along with a sentence describing the results. In this case we
could say:
The results indicated that the model was a significant predictor of exam performance,
F(2,26) = 9.34, p = .001.
Coefficients
While the ANOVA table tells us whether the overall model is a significant predictor of the
outcome variable, this table tells us the extent to which the individual predictor variables
contribute to the model.
There are two sections of the table that you need to look at to interpret your multiple
regression. The first part of the table that we need to look at is the Sig column. This tells us
whether the predictors significantly contributed to the model or not.
By reading across the rows for each of the predictor variables, we can see that:
Remember… this example is fictional!! Please don’t go away thinking that revision does not
lead to better marks!!! ☺
The next column we need to look at contains the unstandardized beta coefficients for the
model (the B values). These values tell us about the relationships between the outcome and
both predictor variables. As both values are positive, so are the relationships. That is, as time
spent revising increases (or as subject enjoyment scores go up), exam scores also get higher. In
addition, these B values give us an idea of the influence each predictor has on the outcome if
the effects of the other variables are held constant.
• Revision Intensity (B1 = .295): as revision intensity increases by one unit (i.e. by one
hour), exam scores increase by 0.295 units.
• Subject Enjoyment (B2 = .668): as people’s enjoyment of the subject increased by one
unit on the enjoyment scale, exam scores went up by 0.668 units.
In regression, we can produce a statistical model that allows us to predict values of our
outcome variable based on our predictor variable. This table also gives us all of the information
we need to do that. This model takes the form of a statistical equation where:
Y = B0 + B1X1 + B2X2
We can also use the B column to create our predictive model. To do this, we need to replace
the Bs with their actual numerical values. B0 is found in the row labelled (Constant); B1 in the
row for the first predictor (Revision Intensity); B2 in the row for the second predictor (Subject
Enjoyment).
Replacing the Bs with the correct values gives us the following predictive model:
By inserting the hours participants spent revising and their subject enjoyment score, we can
now predict how well somebody is likely to do on their exam.
When writing up your results there are certain statistics that you need to report:
• First, you need to state the proportion of variance that can be explained by your model.
This is represented by the statistic R2 and is a number between 0 and 1. It can either be
reported in this format (e.g. R2 = .418) or it can be multiplied by 100 to represent the
percentage of variance your model explains (e.g. 41.8%).
• Second, you need to report whether or not your model was a significant predictor of the
outcome variable using the results of the ANOVA.
• Finally, you need to include information about your predictor variables. In this case, you
need to include your B values for both variables and the significance of their
contribution to the model. It is also a good idea to include your final model here.
Writing up the results
Now you have seen what information you need to include in a results section, for this example
you might write something like this:
A multiple regression was carried out to investigate whether Revision Intensity and
Subject Enjoyment could significantly predict participants’ exam scores. The results of
the regression indicated that the model explained 41.8% of the variance and that the
model was a significant predictor of exam performance, F(2,26) = 9.34, p = .001. While
Subject Enjoyment contributed significantly to the model (B = .668, p<.05), Revision
Intensity did not (B = .295, p=.07). The final predictive model was:
This brings us to the end of the tutorial. Why not download the dataset used in this tutorial and
see if you can produce the same output on your own. Remember, practice makes perfect!