Psychology 282 Lecture #8 Outline Semipartial and Partial Correlations Overall strength of association in MLR with 2 IVs is measured by squared multiple correlation: 2 Y .12 R rY21 + rY22 − 2rY 1rY 2 r12 = 1 − r122 This is the proportion of variance in Y accounted for by its linear relationship with X1 and X2. Consider the problem of determining the contribution of each of the two IVs in accounting for variance in Y. This information would help us understand the relative importance of each IV. As we shall see, there are different ways to measure these contributions statistically, and these measures have different meanings. 2 Representation of relationship among variables in form of Venn diagram: Y e b a c d X2 X1 Squared correlations as represented in diagram: r122 = c + d rY21 = a + c rY22 = b + c RY2.12 = a + b + c Now consider how to define the contribution of one IV; start with X1. What is the unique contribution of X1? Could say that unique contribution of X1 is represented by area a in diagram. 3 This area is called the squared semipartial 2 correlation of X1 with Y: sr1 2 In terms of the diagram: sr1 = a Computing this value: 2 R Squared multiple correlation: Y .12 = a + b + c 2 Variance accounted for by X2: rY 2 = b + c Variance accounted for uniquely by X1 would then be found by: sr12 = RY2.12 − rY22 That is, variance accounted for by X1 above and beyond that accounted for by X2 is obtained by taking total variance accounted for by X1 and X2 and subtracting that portion accounted for by X2. Important concept: This value measures the increase, or increment, in R2 when we add X1 as an IV to a regression model that already contains X2 as an IV. Note the conceptual and statistical difference between the simple contribution of X1 to explaining 2 r the variance of Y (defined by Y 1 ) and the unique contribution of X1 to explaining the variance of Y 2 (defined by sr1 ). 4 Now consider applying same rationale to determine the unique contribution of X2: Diagram: Area b Defines squared semipartial correlation of X2 with Y: sr22 = b Computing this value: 2 Squared multiple correlation: RY .12 = a + b + c 2 Variance accounted for by X1: rY 1 = a + c Variance accounted for uniquely by X2 would then be found by: sr22 = RY2.12 − rY21 That is, variance accounted for by X2 above and beyond that accounted for by X1 is obtained by taking total variance accounted for by X1 and X2 and subtracting that portion accounted for by X1. Important concept: This value measures the increase, or increment, in R2 when we add X2 as an IV to a regression model that already contains X1 as an IV. In applications of MLR it is useful to obtain these sr2 values for each IV. They are interpreted as representing the unique contribution of each IV in 5 accounting for the variance in Y. Equivalently, they represent the increment in variance accounted for by that IV when it is added to the model. 2 2 2 Important: Note that sr1 + sr2 ≠ RY .12 We cannot add the unique contributions of the separate IVs and obtain the total variance accounted for. The total variance accounted includes a portion that is accounted for “redundantly” by the two IVs due to their correlation with each other. That redundant portion is represented by area c in the diagram. It would be tempting to try to define area c 2 2 2 formally as RY .12 − sr1 − sr2 . However, this is not correct because this computation can produce a negative value in some situations. Thus, the redundant contribution of the IVs cannot be calculated. Note also that if r12 is high, the unique contribution of each IV will be small. That is, the squared semipartial correlations will be small. 6 There is another way to define the contribution of each separate IV. Look at the diagram again: Y e b a c d X1 X2 The first approach focused on defining a proportion of the total variance in Y that is accounted for by each IV; this led us to focus on areas a and b as a proportion of the full Y circle. Considering first X1, note that we could define its contribution in terms of a proportion of that part of the Y variance that is not accounted for by X2. The Y variance not accounted for by X2 is represented by area a+e. The portion of that area that is explained by X1 is represented by area a. Taking a proportion, we can define the contribution of X1 as a/(a+e). 7 This quantity is called a squared partial correlation: pr12 = a a+e Consider how this value could be computed. As noted earlier, area a is equivalent to the squared semipartial correlation for X1 with Y: sr12 = RY2.12 − rY22 Area (a+e) would be the variance in Y not explained 2 ( 1 − r by X2, which could be calculated as: Y2) Combining these two values yields: RY2.12 − rY22 pr = 1 − rY22 2 1 This value tells us the following: Of the variance in Y that is not explained by X2, what proportion is explained by X1. Interpretation in terms of increments in R2 is as follows: This value tells us the ratio of the increment in R2 associated with X1 divided by the maximum increment in R2 that could have been achieved. 8 Note that pr1 ≥ sr1 . These values will be equal only 2 when rY 2 = 0 . 2 2 Now we can define the same squared partial correlation to represent the contribution of X2: the proportion of that part of the Y variance that is not accounted for by X1. The Y variance not accounted for by X1 is represented by area b+e. The portion of that area that is explained by X2 is represented by area b. Taking a proportion, we can define the contribution of X2 as b/(b+e). This quantity is called a squared partial correlation: pr22 = b b+e Consider how this value could be computed. As noted earlier, area b is equivalent to the squared semipartial correlation for X2 with Y: sr22 = RY2.12 − rY21 9 Area (b+e) would be the variance in Y not explained 2 by X1, which could be calculated as: (1 − rY 1 ) Combining these two values yields: RY2.12 − rY21 pr = 1 − rY21 2 2 This value tells us the following: Of the variance in Y that is not explained by X1, what proportion is explained by X2. These pr2 values can be obtained for each IV in the regression model. Interpretation in terms of increments in R2 would be as follows: The pr2 value for a particular IV would indicate the ratio of the increment in R2 that is associated with that variable divided by the maximum possible increment that could have been achieved. 10 Defining semipartial and partial correlations by “residualizing” In the developments presented above we defined squared semipartial and squared partial correlations in terms of portions of variance in Y accounted for, or in terms of increments in R2. Another way to define these values is based on the concept of residualizing. Based on residualizing, here is how we would define the semipartial correlation of X1 with Y: First, suppose we were to conduct a SLR analysis treating X1 as a dependent variable and X2 as an independent variable, ignoring Y. Xˆ 1 = B0 + B X1 X 2 X 2 Residuals: e X1 = X 1 − Xˆ 1 These residuals represent that part of X1 that is not accounted for by X2. (X1 has been residualized by removing the effect of X2 from it. Another way to say this is that the effect of X2 on X1 has been partialed out.) 11 Now suppose we compute the correlation between these residuals and variable Y: reX 1Y This correlation is the semipartial correlation of X1 with Y, removing the effect of X2 from X1. That is, sr1 = re X 1Y By writing out the full expression for reX 1Y , then simplifying, one can show that sr1 = re X 1Y = rY 1 − rY 2 r12 1 − r122 Squaring this value then would produce the squared semipartial correlation defined earlier. To define sr2 by residualizing, one would follow exactly the same procedure just shown, but reverse the roles of X1 and X2: SLR predicting X2 from X1: Residuals: Xˆ 2 = B0 + BX 2 X 1 X 1 eX 2 = X 2 − Xˆ 2 These residuals represent that part of X2 that is not accounted for by X1. 12 Obtain correlation between these residuals and Y: sr2 = re X 2Y = rY 2 − rY 1r12 1 − r122 This is the semipartial correlation of X2 with Y, partialing X1 from X2. Squaring this value produces the squared semipartial as presented earlier. We can also use the concept and method of residualizing to define a partial correlation. This is an extension of the procedure just described in that for a partial correlation we partial out the effect of one predictor from the other predictor and also from the dependent variable. To define the partial correlation of X1 with Y, partialing out X2: First partial X2 from X1: SLR: Residuals: Xˆ 1 = B0 + B X1 X 2 X 2 e = X − Xˆ X1 1 1 These residuals represent that part of X1 that is not accounted for by X2. 13 Then partial X2 from Y: Yˆ = B0 + BYX X 2 SLR: 2 Residuals: eY = Y − Yˆ These residuals represent that part of Y that is not accounted for by X2. The partial correlation is then obtained by correlating the two sets of residuals: pr1 = re X 1eY By writing out this correlation algebraically and then simplifying, it can be shown that pr1 = re X 1eY = rY 1 − rY 2 r12 1 − rY22 1 − r122 This is the partial correlation of X1 with Y, partialing out the effect of X2. Squaring this value produces the squared partial correlation covered earlier, and the corresponding interpretation. 14 To define the partial correlation for X2 with Y, partialing out X1, we follow exactly the same approach but reverse the roles of X1 and X2. First partial X1 from X2: SLR: Residuals: Xˆ 2 = B0 + BX 1 X 2 X 1 e = X − Xˆ X2 2 2 These residuals represent that part of X2 that is not accounted for by X1. Then partial X1 from Y: Yˆ = B0 + BYX X 1 SLR: 1 Residuals: eY = Y − Yˆ These residuals represent that part of Y that is not accounted for by X1. The partial correlation is then obtained by correlating the two sets of residuals: pr2 = re X 2eY 15 By writing out this correlation algebraically and the simplifying, it can be shown that pr2 = re X 2eY = rY 2 − rY 1r12 1 − rY21 1 − r122 This is the partial correlation of X2 with Y, partialing out the effect of X1. The concept and methods of residualizing are important in correlational methods. They can be used in many situations when one wishes to statistically control for effects of some variables on other variables. A relationship among the various partial coefficients We use the term “partial coefficients” to refer to any of the various partial regression coefficients and correlations we have defined (including raw and standardized regression coefficients, and semipartial and partial correlations). Consider the partial coefficients associated with X1. If we look back at our definitional formulas for the standardized regression weight (β1), the semipartial correlation (sr1), and the partial correlation (pr1), we 16 can see that all of these formulas have the same numerator: (rY1 – rY2r12). This observation has an important implication: If any one of these values is zero, then they all must be zero. This relationship extends to the raw score regression weight (B1), since there is a simple multiplicative relationship between B1 and β1. Thus, in considering any single IV, if any of the partial coefficients representing the effect of that IV are zero, then all are zero. This fact has useful consequences in the context of testing hypotheses about partial coefficients, which will be covered when we study inferences in multiple regression.
© Copyright 2026 Paperzz