Multiple Correlation Coefficient PDF
Multiple Correlation Coefficient PDF
Multiple Correlation Coefficient PDF
Herv Abdi1
1 Overview
The multiple correlation coefficient generalizes the standard coefficient of correlation. It is used in multiple regression analysis to
assess the quality of the prediction of the dependent variable. It
corresponds to the squared correlation between the predicted and
the actual values of the dependent variable. It can also be interpreted as the proportion of the variance of the dependent variable
explained by the independent variables. When the independent
variables (used for predicting the dependent variable) are pairwise
orthogonal, the multiple correlation coefficient is equal to the sum
of the squared coefficients of correlation between each independent variable and the dependent variable. This relation does not
hold when the independent variables are not orthogonal. The significance of a multiple coefficient of correlation can be assessed
with an F ratio. The magnitude of the multiple coefficient of correlation tends to overestimate the magnitude of the population correlation, but it is possible to correct for this overestimation. Strictly
speaking we should refer to this coefficient as the squared multiple correlation coefficient, but current usage seems to ignore the
1
(1)
1 x 1,1
..
..
.
.
X = 1 x n,1
..
..
.
.
1 x N ,1
..
.
..
.
x 1, j
.. . .
.
.
x n, j
.. . .
.
.
xN , j
x 1,J
..
.
x n,J
..
.
x N ,J
y1
..
.
y
and y =
n
..
.
yN
(2)
with
T
XT y .
b= X X
(3)
1 T 2
(1 y)
N
(4)
1 T 2
(1 y) .
N
(5)
(6)
SS regression
SS regression + SS error
SS regression
SS total
(7)
R Y2 .1,...,J
1 R Y2 .1,...,J
N J 1
.
J
(8)
Under the usual assumptions of normality of the error and of independence of the error and the scores, this F ratio is distributed
under the null hypothesis as a Fisher distribution with 1 = J and
2 = N J 1 degrees of freedom.
35
39
40
52
61
73
21
31
34
42
58
66
6
8
18
26
46
52
N 1
2
2
.
(9)
ReY .1,...,J = 1 1 R Y .1,...,J
N J 1
3 Example 1:
Multiple correlation coefficient
with orthogonal predictors
When the independent variables are pairwise orthogonal, the importance of each of them in the regression is assessed by computing the squared coefficient of correlation between each of the independent variables and the dependent variable. The sum of these
squared coefficients of correlation is equal to the multiple coefficient of correlation. We illustrate this case with the data from Table 1. In this example, the dependent variable (Y ) is the number or
sentences recalled by participants who learned a list of unrelated
sentences. The first independent variable or first predictor, X 1 is
the number of trials used to learn the list. It takes the values 2,
4, and 8. It is expected that recall will increase as a function of the
number of trials. The second independent variable, X 2 is the number of additional interpolated lists that the participants are asked
to learned. It takes the values 2, 4, and 8. As a consequence of
retroactive inhibition, it is expected that recall will decrease as a
function of the number of interpolated lists learned.
Using Equation 3, we found that Yb can be obtained from X 1
and X 2 as
Yb = 30 + 6 X 1 4 X 2 .
(10)
Using these data and Equations 4 and 5, we find that
SS regression = 5824,
SS total = 6214,
(11)
This gives the following value for the multiple coefficient of correlation:
SS regression 5824
=
= .9372 .
(12)
R Y2 .1,...,J =
SS total
6214
In order to decide if this value of R Y2 .1,...,J is large enough to be considered significant, we compute an F ratio equal to
F=
R Y2 .1,...,J
1 R Y2 .1,...,J
N J 1
.9372
15
=
= 111.93 .
J
1 .9372 2
5
(13)
(14)
A better estimate of the population value of the multiple coefficient of correlation can obtained as
N 1
17
2
2
e
= 1(1.9372) = .9289 .
R Y .1,...,J = 1 1 R Y .1,...,J
N J 1
15
(15)
4 Example 2:
Multiple correlation coefficient
with non-orthogonal predictors
When the independent variables are correlated, the multiple coefficient of correlation is not equal to the sum of the squared correlation coefficients between the dependent variable and the independent variables. In fact, such a strategy would overestimate the
contribution of each variable because the variance that they share
would be counted several times.
For example, consider the data given in Table 2 where the dependent variable is to be predicted from the independent variables
X 1 and X 2 . The prediction of the dependent variable (using Equation 3) is found to be equal to
Yb = 1.67 + X 1 + 9.50X 2 ;
(16)
the child can pronounce in a given time). Six children were tested.
Y (Memory span)
X 1 (age)
X 2 (Speech rate)
14
4
1
23
4
2
30
7
2
50
7
4
39
10
3
67
10
6
(17)
SS total = 1846.83,
R Y2 .1,...,J
1 R Y2 .1,...,J
N J 1
.9866
3
=
= 110.50 .
J
1 .9866 2
(20)
N 1
5
2
2
e
R Y .1,...,J = 1 1 R Y .1,...,J
= 1 (1 .9866) = .9776 .
N J 1
2
(21)
7
References
[1] Abdi, H., Dowling, W.J., Valentin, D., Edelman, B., & Posamentier M. (2002). Experimental Design and research methods. Unpublished manuscript. Richardson: The University of
Texas at Dallas, Program in Cognition.
[2] Cohen, J., & Cohen, P. (1983). Applied multiple regression/correlation analysis for the behavioral sciences (2nd edition). Hillsdale (NJ): Erlbaum.
[3] Darlington, R.B. (1990). Regression and linear models. NewYork:
McGraw-Hill.
[4] Pedhazur, E.J. (1997). Multiple regression in behavioral research. (3rd edition) New York: Holt, Rinehart and Winston,
Inc.