ZIKMUND BABIN CARR GRIFFIN BUSINESS MARKET RESEARCH EIGHTH EDITION © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Chapter 23 Bivariate Statistical Analysis: Measures of Association LEARNING OUTCOMES After studying this chapter, you should be able to 1. Apply and interpret simple bivariate correlations 2. Interpret a correlation matrix 3. Understand simple (bivariate) regression 4. Understand the least-squares estimation technique 5. Interpret regression output including the tests of hypotheses tied to specific parameter coefficients © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–2 The Basics • Measures of Association Refers to a number of bivariate statistical techniques used to measure the strength of a relationship between two variables. The chi-square (2) test provides information about whether two or more less-than interval variables are interrelated. Correlation analysis is most appropriate for interval or ratio variables. Regression can accommodate either less-than interval or interval independent variables, but the dependent variable must be continuous. © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–3 EXHIBIT 23.1 Bivariate Analysis— Common Procedures for Testing Association © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–4 Simple Correlation Coefficient (continued) • Correlation coefficient A statistical measure of the covariation, or association, between two at-least interval variables. • Covariance Extent to which two variables are associated systematically with each other. X i X Yi Y n rxy ryx i 1 Xi X Yi Y n i 1 © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 2 n 2 i 1 23–5 Simple Correlation Coefficient • Correlation coefficient (r) Ranges from +1 to -1 Perfect positive linear relationship = +1 Perfect negative (inverse) linear relationship = -1 No correlation = 0 • Correlation coefficient for two variables (X,Y) © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–6 EXHIBIT 23.2 Scatter Diagram to Illustrate Correlation Patterns © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–7 Correlation, Covariance, and Causation • When two variables covary, they display concomitant variation. • This systematic covariation does not in and of itself establish causality. • e.g., Rooster’s crow and the rising of the sun Rooster does not cause the sun to rise. © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–8 EXHIBIT 23.3 Correlation Analysis of Number of Hours Worked in Manufacturing Industries with Unemployment Rate © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–9 Coefficient of Determination • Coefficient of Determination (R2) A measure obtained by squaring the correlation coefficient; the proportion of the total variance of a variable accounted for by another value of another variable. Measures that part of the total variance of Y that is accounted for by knowing the value of X. Explained variance R Total Variance 2 © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–10 Correlation Matrix • Correlation matrix The standard form for reporting correlation coefficients for more than two variables. • Statistical Significance The procedure for determining statistical significance is the t-test of the significance of a correlation coefficient. © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–11 EXHIBIT 23.4 Pearson Product-Moment Correlation Matrix for Salesperson Examplea © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–12 Regression Analysis • Simple (Bivariate) Linear Regression A measure of linear association that investigates straight-line relationships between a continuous dependent variable and an independent variable that is usually continuous, but can be a categorical dummy variable. • The Regression Equation (Y = α + βX ) Y = the continuous dependent variable X = the independent variable α = the Y intercept (regression line intercepts Y axis) β = the slope of the coefficient (rise over run) © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–13 Regression Line and Slope Y 130 120 110 100 Yˆ aˆ ̂X 90 X 80 Yˆ X 80 90 100 110 © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 120 130 140 150 160 170 23–14 The Regression Equation • Parameter Estimate Choices β is indicative of the strength and direction of the relationship between the independent and dependent variable. α (Y intercept) is a fixed point that is considered a constant (how much Y can exist without X) • Standardized Regression Coefficient (β) Estimated coefficient of the strength of relationship between the independent and dependent variables. Expressed on a standardized scale where higher absolute values indicate stronger relationships (range is from -1 to 1). © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–15 The Regression Equation (cont’d) • Parameter Estimate Choices Raw regression estimates (b1) Raw regression weights have the advantage of retaining the scale metric—which is also their key disadvantage. If the purpose of the regression analysis is forecasting, then raw parameter estimates must be used. This is another way of saying when the researcher is interested only in prediction. Standardized regression estimates (β) Standardized regression estimates have the advantage of a constant scale. Standardized regression estimates should be used when the researcher is testing explanatory hypotheses. © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–16 EXHIBIT 23.5 The Advantage of Standardized Regression Weights © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–17 EXHIBIT 23.6 Relationship of Sales Potential to Building Permits Issued © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–18 EXHIBIT 23.7 The Best Fit Line or Knocking Out the Pins © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–19 Ordinary Least-Squares (OLS) Method of Regression Analysis • OLS Guarantees that the resulting straight line will produce the least possible total error in using X to predict Y. Generates a straight line that minimizes the sum of squared deviations of the actual values from this predicted regression line. No straight line can completely represent every dot in the scatter diagram. There will be a discrepancy between most of the actual scores (each dot) and the predicted score . Uses the criterion of attempting to make the least amount of total error in prediction of Y from X. © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–20 Ordinary Least-Squares Method of Regression Analysis (OLS) (cont’d) © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–21 Ordinary Least-Squares Method of Regression Analysis (OLS) (cont’d) The equation means that the predicted value for any value of X (Xi) is determined as a function of the estimated slope coefficient, plus the estimated intercept coefficient + some error. © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–22 Ordinary Least-Squares Method of Regression Analysis (OLS) (cont’d) © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–23 Ordinary Least-Squares Method of Regression Analysis (OLS) (cont’d) • Statistical Significance Of Regression Model • F-test (regression) Determines whether more variability is explained by the regression or unexplained by the regression. © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–24 Ordinary Least-Squares Method of Regression Analysis (OLS) (cont’d) • Statistical Significance Of Regression Model ANOVA Table: © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–25 Ordinary Least-Squares Method of Regression Analysis (OLS) (cont’d) • R2 The proportion of variance in Y that is explained by X (or vice versa) A measure obtained by squaring the correlation coefficient; that proportion of the total variance of a variable that is accounted for by knowing the value of another variable. 3,398.49 R 0.875 3,882.40 2 © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–26 EXHIBIT 23.8 Simple Regression Results for Building Permit Example © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–27 EXHIBIT 23.9 OLS Regression Line © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–28 Simple Regression and Hypothesis Testing • The explanatory power of regression lies in hypothesis testing. Regression is often used to test relational hypotheses. The outcome of the hypothesis test involves two conditions that must both be satisfied: The regression weight must be in the hypothesized direction. Positive relationships require a positive coefficient and negative relationships require a negative coefficient. The t-test associated with the regression weight must be significant. © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–29 EXHIBIT 23A.1 Least-Squares Computation © 2010 South-Western/Cengage Learning. All rights reserved. May not be scanned, copied or duplicated, or posted to a publically accessible website, in whole or in part. 23–30
© Copyright 2026 Paperzz