Chapter 15 Multiple Regression n Multiple Regression Model

  • Slides: 65
Download presentation
Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple

Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple Coefficient of Determination n Model Assumptions n Testing for Significance n Using the Estimated Regression Equation for Estimation and Prediction n Categorical Independent Variables n Residual Analysis n Logistic Regression © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 1

Multiple Regression n n In this chapter we continue our study of regression analysis

Multiple Regression n n In this chapter we continue our study of regression analysis by considering situations involving two or more independent variables. This subject area, called multiple regression analysis, enables us to consider more factors and thus obtain better estimates than are possible with simple linear regression. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 2

Multiple Regression Model n Multiple Regression Model The equation that describes how the dependent

Multiple Regression Model n Multiple Regression Model The equation that describes how the dependent variable y is related to the independent variables x 1, x 2, . . . xp and an error term is: y = 0 + 1 x 1 + 2 x 2 +. . . + pxp + where: 0, 1, 2, . . . , p are the parameters, and is a random variable called the error term © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 3

Multiple Regression Equation n Multiple Regression Equation The equation that describes how the mean

Multiple Regression Equation n Multiple Regression Equation The equation that describes how the mean value of y is related to x 1, x 2, . . . xp is: E(y) = 0 + 1 x 1 + 2 x 2 +. . . + pxp © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 4

Estimated Multiple Regression Equation n Estimated Multiple Regression Equation ^ y y = b

Estimated Multiple Regression Equation n Estimated Multiple Regression Equation ^ y y = b 0 + b 1 x 1 + b 2 x 2 +. . . + bpxp A simple random sample is used to compute sample statistics b 0, b 1, b 2, . . . , bp that are used as the point estimators of the parameters 0, 1, 2, . . . , p. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 5

Estimation Process Multiple Regression Model E(y) = 0 + 1 x 1 + 2

Estimation Process Multiple Regression Model E(y) = 0 + 1 x 1 + 2 x 2 +. . . + pxp + Multiple Regression Equation E(y) = 0 + 1 x 1 + 2 x 2 +. . . + pxp Unknown parameters are Sample Data: x 1 x 2. . . xp y. . . . 0, 1, 2, . . . , p b 0, b 1, b 2, . . . , bp Estimated Multiple Regression Equation provide estimates of 0, 1, 2, . . . , p Sample statistics are b 0, b 1, b 2, . . . , bp © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 6

Least Squares Method n Least Squares Criterion n Computation of Coefficient Values The formulas

Least Squares Method n Least Squares Criterion n Computation of Coefficient Values The formulas for the regression coefficients b 0, b 1, b 2, . . . bp involve the use of matrix algebra. We will rely on computer software packages to perform the calculations. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 7

Least Squares Method n Computation of Coefficient Values The formulas for the regression coefficients

Least Squares Method n Computation of Coefficient Values The formulas for the regression coefficients b 0, b 1, b 2, . . . bp involve the use of matrix algebra. We will rely on computer software packages to perform the calculations. The emphasis will be on how to interpret the computer output rather than on how to make the multiple regression computations. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 8

Multiple Regression Model n Example: Programmer Salary Survey A software firm collected data for

Multiple Regression Model n Example: Programmer Salary Survey A software firm collected data for a sample of 20 computer programmers. A suggestion was made that regression analysis could be used to determine if salary was related to the years of experience and the score on the firm’s programmer aptitude test. The years of experience, score on the aptitude test, and corresponding annual salary ($1000 s) for a sample of 20 programmers is shown on the next slide. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 9

Multiple Regression Model Exper. (Yrs. ) Test Score Salary ($000 s) 4 7 1

Multiple Regression Model Exper. (Yrs. ) Test Score Salary ($000 s) 4 7 1 5 8 10 0 1 6 6 78 100 86 82 86 84 75 80 83 91 24. 0 43. 0 23. 7 34. 3 35. 8 38. 0 22. 2 23. 1 30. 0 33. 0 9 2 10 5 6 8 4 6 3 3 88 73 75 81 74 87 79 94 70 89 38. 0 26. 6 36. 2 31. 6 29. 0 34. 0 30. 1 33. 9 28. 2 30. 0 © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 10

Multiple Regression Model Suppose we believe that salary (y) is related to the years

Multiple Regression Model Suppose we believe that salary (y) is related to the years of experience (x 1) and the score on the programmer aptitude test (x 2) by the following regression model: y = 0 + 1 x 1 + 2 x 2 + where y x 1 x 2 = annual salary ($000) = years of experience = score on programmer aptitude test © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 11

Solving for the Estimates of 0, 1, 2 Least Squares Output Input Data x

Solving for the Estimates of 0, 1, 2 Least Squares Output Input Data x 1 x 2 y 4 78 24 7 100 43. . . 3 89 30 Computer Package for Solving Multiple Regression Problems © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. b 0 = b 1 = b 2 = R 2 = etc. Slide 12

Solving for the Estimates of 0, 1, 2 n Regression Equation Output p Predictor

Solving for the Estimates of 0, 1, 2 n Regression Equation Output p Predictor Coef SE Coef T Constant 3. 17394 6. 15607 0. 5156 0. 61279 Experience 1. 4039 0. 19857 7. 0702 1. 9 E-06 Test Score 0. 25089 0. 07735 3. 2433 0. 00478 © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 13

Estimated Regression Equation SALARY = 3. 174 + 1. 404(EXPER) + 0. 251(SCORE) Note:

Estimated Regression Equation SALARY = 3. 174 + 1. 404(EXPER) + 0. 251(SCORE) Note: Predicted salary will be in thousands of dollars. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 14

Interpreting the Coefficients In multiple regression analysis, we interpret each regression coefficient as follows:

Interpreting the Coefficients In multiple regression analysis, we interpret each regression coefficient as follows: bi represents an estimate of the change in y corresponding to a 1 -unit increase in xi when all other independent variables are held constant. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 15

Interpreting the Coefficients b 1 = 1. 404 Salary is expected to increase by

Interpreting the Coefficients b 1 = 1. 404 Salary is expected to increase by $1, 404 for each additional year of experience (when the variable score on programmer attitude test is held constant). © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 16

Interpreting the Coefficients b 2 = 0. 251 Salary is expected to increase by

Interpreting the Coefficients b 2 = 0. 251 Salary is expected to increase by $251 for each additional point scored on the programmer aptitude test (when the variable years of experience is held constant). © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 17

Multiple Coefficient of Determination n Relationship Among SST, SSR, SSE SST = SSR +

Multiple Coefficient of Determination n Relationship Among SST, SSR, SSE SST = SSR + SSE = + where: SST = total sum of squares SSR = sum of squares due to regression SSE = sum of squares due to error © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 18

Multiple Coefficient of Determination n ANOVA Output Analysis of Variance SOURCE Regression Residual Error

Multiple Coefficient of Determination n ANOVA Output Analysis of Variance SOURCE Regression Residual Error Total DF 2 17 19 SST SS 500. 3285 99. 45697 599. 7855 MS 250. 164 5. 850 F 42. 76 P 0. 000 SSR © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 19

Multiple Coefficient of Determination R 2 = SSR/SST R 2 = 500. 3285/599. 7855

Multiple Coefficient of Determination R 2 = SSR/SST R 2 = 500. 3285/599. 7855 =. 83418 © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 20

Adjusted Multiple Coefficient of Determination n Adding independent variables, even ones that are not

Adjusted Multiple Coefficient of Determination n Adding independent variables, even ones that are not statistically significant, causes the prediction errors to become smaller, thus reducing the sum of squares due to error, SSE. Because SSR = SST – SSE, when SSE becomes smaller, SSR becomes larger, causing R 2 = SSR/SST to increase. The adjusted multiple coefficient of determination compensates for the number of independent variables in the model. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 21

Adjusted Multiple Coefficient of Determination © 2014 Cengage Learning. All Rights Reserved. May not

Adjusted Multiple Coefficient of Determination © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 22

Assumptions About the Error Term The error is a random variable with mean of

Assumptions About the Error Term The error is a random variable with mean of zero. The variance of , denoted by 2, is the same for all values of the independent variables. The values of are independent. The error is a normally distributed random variable reflecting the deviation between the y value and the expected value of y given by 0 + 1 x 1 + 2 x 2 +. . + pxp. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 23

Testing for Significance In simple linear regression, the F and t tests provide the

Testing for Significance In simple linear regression, the F and t tests provide the same conclusion. In multiple regression, the F and t tests have different purposes. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 24

Testing for Significance: F Test The F test is used to determine whether a

Testing for Significance: F Test The F test is used to determine whether a significant relationship exists between the dependent variable and the set of all the independent variables. The F test is referred to as the test for overall significance. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 25

Testing for Significance: t Test If the F test shows an overall significance, the

Testing for Significance: t Test If the F test shows an overall significance, the t test is used to determine whether each of the individual independent variables is significant. A separate t test is conducted for each of the independent variables in the model. We refer to each of these t tests as a test for individual significance. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 26

Testing for Significance: F Test Hypotheses H 0: 1 = 2 =. . .

Testing for Significance: F Test Hypotheses H 0: 1 = 2 =. . . = p = 0 Ha: One or more of the parameters is not equal to zero. Test Statistics F = MSR/MSE Rejection Rule Reject H 0 if p-value < a or if F > F , where F is based on an F distribution with p d. f. in the numerator and n - p - 1 d. f. in the denominator. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 27

F Test for Overall Significance Hypotheses Rejection Rule H 0: 1 = 2 =

F Test for Overall Significance Hypotheses Rejection Rule H 0: 1 = 2 = 0 Ha: One or both of the parameters is not equal to zero. For =. 05 and d. f. = 2, 17; F. 05 = 3. 59 Reject H 0 if p-value <. 05 or F > 3. 59 © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 28

F Test for Overall Significance n ANOVA Output Analysis of Variance SOURCE Regression Residual

F Test for Overall Significance n ANOVA Output Analysis of Variance SOURCE Regression Residual Error Total DF 2 17 19 SS 500. 3285 99. 45697 599. 7855 MS 250. 164 5. 850 F 42. 76 P 0. 000 p-value used to test for overall significance © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 29

F Test for Overall Significance Test Statistics Conclusion F = MSR/MSE = 250. 16/5.

F Test for Overall Significance Test Statistics Conclusion F = MSR/MSE = 250. 16/5. 85 = 42. 76 p-value <. 05, so we can reject H 0. (Also, F = 42. 76 > 3. 59) © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 30

Testing for Significance: t Test Hypotheses Test Statistics Rejection Rule Reject H 0 if

Testing for Significance: t Test Hypotheses Test Statistics Rejection Rule Reject H 0 if p-value < a or if t < -t or t > t where t is based on a t distribution with n - p - 1 degrees of freedom. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 31

t Test for Significance of Individual Parameters Hypotheses Rejection Rule For =. 05 and

t Test for Significance of Individual Parameters Hypotheses Rejection Rule For =. 05 and d. f. = 17, t. 025 = 2. 11 Reject H 0 if p-value <. 05, or if t < -2. 11 or t > 2. 11 © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 32

t Test for Significance of Individual Parameters n Regression Equation Output p Predictor Coef

t Test for Significance of Individual Parameters n Regression Equation Output p Predictor Coef SE Coef T Constant 3. 17394 6. 15607 0. 5156 0. 61279 Experience 1. 4039 0. 19857 7. 0702 1. 9 E-06 Test Score 0. 25089 0. 07735 3. 2433 0. 00478 t statistic and p-value used to test for the individual significance of “Experience” © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 33

t Test for Significance of Individual Parameters Test Statistics Conclusions Reject both H 0:

t Test for Significance of Individual Parameters Test Statistics Conclusions Reject both H 0: 1 = 0 and H 0: 2 = 0. Both independent variables are significant. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 34

Testing for Significance: Multicollinearity The term multicollinearity refers to the correlation among the independent

Testing for Significance: Multicollinearity The term multicollinearity refers to the correlation among the independent variables. When the independent variables are highly correlated (say, |r | >. 7), it is not possible to determine the separate effect of any particular independent variable on the dependent variable. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 35

Testing for Significance: Multicollinearity If the estimated regression equation is to be used only

Testing for Significance: Multicollinearity If the estimated regression equation is to be used only for predictive purposes, multicollinearity is usually not a serious problem. Every attempt should be made to avoid including independent variables that are highly correlated. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 36

Using the Estimated Regression Equation for Estimation and Prediction The procedures for estimating the

Using the Estimated Regression Equation for Estimation and Prediction The procedures for estimating the mean value of y and predicting an individual value of y in multiple regression are similar to those in simple regression. We substitute the given values of x 1, x 2, . . . , xp into the estimated regression equation and use the corresponding value of y as the point estimate. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 37

Using the Estimated Regression Equation for Estimation and Prediction The formulas required to develop

Using the Estimated Regression Equation for Estimation and Prediction The formulas required to develop interval estimates for the mean value of y ^and for an individual value of y are beyond the scope of the textbook. Software packages for multiple regression will often provide these interval estimates. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 38

Categorical Independent Variables In many situations we must work with categorical independent variables such

Categorical Independent Variables In many situations we must work with categorical independent variables such as gender (male, female), method of payment (cash, check, credit card), etc. For example, x 2 might represent gender where x 2 = 0 indicates male and x 2 = 1 indicates female. In this case, x 2 is called a dummy or indicator variable. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 39

Categorical Independent Variables n Example: Programmer Salary Survey As an extension of the problem

Categorical Independent Variables n Example: Programmer Salary Survey As an extension of the problem involving the computer programmer salary survey, suppose that management also believes that the annual salary is related to whether the individual has a graduate degree in computer science or information systems. The years of experience, the score on the programmer aptitude test, whether the individual has a relevant graduate degree, and the annual salary ($000) for each of the sampled 20 programmers are shown on the next slide. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 40

Categorical Independent Variables Exper. Test Salary (Yrs. ) Score Degr. ($000 s) 4 7

Categorical Independent Variables Exper. Test Salary (Yrs. ) Score Degr. ($000 s) 4 7 1 5 8 10 0 1 6 6 78 100 86 82 86 84 75 80 83 91 No Yes Yes No No No Yes 24. 0 43. 0 23. 7 34. 3 35. 8 38. 0 22. 2 23. 1 30. 0 33. 0 Exper. Test Salary (Yrs. ) Score Degr. ($000 s) 9 2 10 5 6 8 4 6 3 3 88 73 75 81 74 87 79 94 70 89 © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Yes No No 38. 0 26. 6 36. 2 31. 6 29. 0 34. 0 30. 1 33. 9 28. 2 30. 0 Slide 41

Estimated Regression Equation ^ y y = b 0 + b 1 x 1

Estimated Regression Equation ^ y y = b 0 + b 1 x 1 + b 2 x 2 + b 3 x 3 where: y^ = annual salary ($1000) x 1 = years of experience x 2 = score on programmer aptitude test x 3 = 0 if individual does not have a graduate degree 1 if individual does have a graduate degree x 3 is a dummy variable © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 42

Categorical Independent Variables n ANOVA Output Analysis of Variance SOURCE Regression Residual Error Total

Categorical Independent Variables n ANOVA Output Analysis of Variance SOURCE Regression Residual Error Total DF 3 16 19 SS 507. 8960 91. 8895 599. 7855 MS 269. 299 5. 743 F 29. 48 P 0. 000 Previously, R Square =. 8342 R 2 = 507. 896/599. 7855 =. 8468 © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Previously, Adjusted R Square =. 815 Slide 43

Categorical Independent Variables n Regression Equation Output Predictor Constant Experience Test Score Grad. Degr.

Categorical Independent Variables n Regression Equation Output Predictor Constant Experience Test Score Grad. Degr. Coef 7. 945 1. 148 0. 197 2. 280 SE Coef 7. 382 0. 298 0. 090 1. 987 T 1. 076 3. 856 2. 191 1. 148 p 0. 298 0. 001 0. 044 0. 268 Not significant © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 44

More Complex Categorical Variables If a categorical variable has k levels, k - 1

More Complex Categorical Variables If a categorical variable has k levels, k - 1 dummy variables are required, with each dummy variable being coded as 0 or 1. For example, a variable with levels A, B, and C could be represented by x 1 and x 2 values of (0, 0) for A, (1, 0) for B, and (0, 1) for C. Care must be taken in defining and interpreting the dummy variables. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 45

More Complex Categorical Variables For example, a variable indicating level of education could be

More Complex Categorical Variables For example, a variable indicating level of education could be represented by x 1 and x 2 values as follows: Highest Degree Bachelor’s Master’s Ph. D. x 1 x 2 0 1 0 0 0 1 © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 46

Residual Analysis n n For simple linear regression the residual plot against and the

Residual Analysis n n For simple linear regression the residual plot against and the residual plot against x provide the same information. In multiple regression analysis it is preferable to use the residual plot against to determine if the model assumptions are satisfied. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 47

Standardized Residual Plot Against n n n Standardized residuals are frequently used in residual

Standardized Residual Plot Against n n n Standardized residuals are frequently used in residual plots for purposes of: • Identifying outliers (typically, standardized residuals < 2 or > +2) • Providing insight about the assumption that the error term has a normal distribution The computation of the standardized residuals in multiple regression analysis is too complex to be done by hand Excel’s Regression tool can be used © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 48

Standardized Residual Plot Against n Residual Output Observation 1 2 3 4 5 Predicted

Standardized Residual Plot Against n Residual Output Observation 1 2 3 4 5 Predicted Y 27. 89626 37. 95204 26. 02901 32. 11201 36. 34251 Residuals -3. 89626 5. 047957 -2. 32901 2. 187986 -0. 54251 Standard Residuals -1. 771707 2. 295406 -1. 059048 0. 994921 -0. 246689 © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 49

Standardized Residual Plot Against n Standardized Residual Plot Outlier Standardized Residual Plot 3 Standard

Standardized Residual Plot Against n Standardized Residual Plot Outlier Standardized Residual Plot 3 Standard Residuals 2 1 0 -1 0 10 20 30 40 50 -2 -3 Predicted Salary © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 50

Logistic Regression n In many ways logistic regression is like ordinary regression. It requires

Logistic Regression n In many ways logistic regression is like ordinary regression. It requires a dependent variable, y, and one or more independent variables. n Logistic regression can be used to model situations in which the dependent variable, y, may only assume two discrete values, such as 0 and 1. n The ordinary multiple regression model is not applicable. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 51

Logistic Regression n Logistic Regression Equation The relationship between E(y) and x 1, x

Logistic Regression n Logistic Regression Equation The relationship between E(y) and x 1, x 2, . . . , xp is better described by the following nonlinear equation. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 52

Logistic Regression n Interpretation of E(y) as a Probability in Logistic Regression If the

Logistic Regression n Interpretation of E(y) as a Probability in Logistic Regression If the two values of y are coded as 0 or 1, the value of E(y) provides the probability that y = 1 given a particular set of values for x 1, x 2, . . . , xp. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 53

Logistic Regression n Estimated Logistic Regression Equation A simple random sample is used to

Logistic Regression n Estimated Logistic Regression Equation A simple random sample is used to compute sample statistics b 0, b 1, b 2, . . . , bp that are used as the point estimators of the parameters 0, 1, 2, . . . , p. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 54

Logistic Regression n Example: Simmons Stores Simmons’ catalogs are expensive and Simmons would like

Logistic Regression n Example: Simmons Stores Simmons’ catalogs are expensive and Simmons would like to send them to only those customers who have the highest probability of making a $200 purchase using the discount coupon included in the catalog. Simmons’ management thinks that annual spending at Simmons Stores and whether a customer has a Simmons credit card are two variables that might be helpful in predicting whether a customer who receives the catalog will use the coupon to make a $200 purchase. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 55

Logistic Regression n Example: Simmons Stores Simmons conducted a study by sending out 100

Logistic Regression n Example: Simmons Stores Simmons conducted a study by sending out 100 catalogs, 50 to customers who have a Simmons credit card and 50 to customers who do not have the card. At the end of the test period, Simmons noted for each of the 100 customers: 1) the amount the customer spent last year at Simmons, 2) whether the customer had a Simmons credit card, and 3) whether the customer made a $200 purchase. A portion of the test data is shown on the next slide. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 56

Logistic Regression n Simmons Test Data (partial) x 1 x 2 y Customer Annual

Logistic Regression n Simmons Test Data (partial) x 1 x 2 y Customer Annual Spending ($1000) Simmons Credit Card $200 Purchase 1 2 3 4 5 6 7 8 9 10 2. 291 3. 215 2. 135 3. 924 2. 528 2. 473 2. 384 7. 076 1. 182 3. 345 1 1 1 0 0 0 1 0 0 1 0 © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 57

Logistic Regression n Simmons Logistic Regression Table (using Minitab) Predictor Constant Spending Card Coef

Logistic Regression n Simmons Logistic Regression Table (using Minitab) Predictor Constant Spending Card Coef SE Coef Z p -2. 1464 0. 3416 1. 0987 0. 5772 0. 1287 0. 4447 -3. 72 2. 66 2. 47 0. 000 0. 008 0. 013 Odds Ratio 95% CI Lower Upper 1. 41 1. 09 1. 25 3. 00 1. 81 7. 17 Log-Likelihood = -60. 487 Test that all slopes are zero: G = 13. 628, DF = 2, P-Value = 0. 001 © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 58

Logistic Regression n Simmons Estimated Logistic Regression Equation © 2014 Cengage Learning. All Rights

Logistic Regression n Simmons Estimated Logistic Regression Equation © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 59

Logistic Regression n Using the Estimated Logistic Regression Equation • For customers that spend

Logistic Regression n Using the Estimated Logistic Regression Equation • For customers that spend $2000 annually and do not have a Simmons credit card: • For customers that spend $2000 annually and do have a Simmons credit card: © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 60

Logistic Regression n Testing for Significance Hypotheses H 0: 1 = 2 = 0

Logistic Regression n Testing for Significance Hypotheses H 0: 1 = 2 = 0 Ha: One or both of the parameters is not equal to zero. Test Statistics z = bi/sb i Rejection Rule Reject H 0 if p-value < a i © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 61

Logistic Regression n Testing for Significance Conclusions For independent variable x 1: z =

Logistic Regression n Testing for Significance Conclusions For independent variable x 1: z = 2. 66 and the p-value =. 008. Hence, 1 = 0. In other words, x 1 is statistically significant. For independent variable x 2: z = 2. 47 and the p-value =. 013. Hence, 2 = 0. In other words, x 2 is also statistically significant. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 62

Logistic Regression n Odds in Favor of an Event Occurring n Odds Ratio ©

Logistic Regression n Odds in Favor of an Event Occurring n Odds Ratio © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 63

Logistic Regression n Estimated Probabilities Annual Spending $1000 $2000 $3000 $4000 $5000 $6000 $7000

Logistic Regression n Estimated Probabilities Annual Spending $1000 $2000 $3000 $4000 $5000 $6000 $7000 Credit Yes Card No 0. 3305 0. 4099 0. 4943 0. 5791 0. 6594 0. 7315 0. 7931 0. 1413 0. 1880 0. 2457 0. 3144 0. 3922 0. 4759 0. 5610 Computed earlier © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 64

Logistic Regression n Comparing Odds Suppose we want to compare the odds of making

Logistic Regression n Comparing Odds Suppose we want to compare the odds of making a $200 purchase for customers who spend $2000 annually and have a Simmons credit card to the odds of making a $200 purchase for customers who spend $2000 annually and do not have a Simmons credit card. © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part. Slide 65