Multiple Logistic Regression STAT E-150 Statistical Methods.

Slides:



Advertisements
Similar presentations
Chapter 12 Inference for Linear Regression
Advertisements

Regression analysis Linear regression Logistic regression.
Inference for Regression
Logistic Regression.
Simple Logistic Regression
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 13 Nonlinear and Multiple Regression.
Regression With Categorical Variables. Overview Regression with Categorical Predictors Logistic Regression.
April 25 Exam April 27 (bring calculator with exp) Cox-Regression
Chapter 13 Multiple Regression
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
Multipe and non-linear regression. What is what? Regression: One variable is considered dependent on the other(s) Correlation: No variables are considered.
Statistics for Managers Using Microsoft® Excel 5th Edition
Chapter 12 Multiple Regression
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 11 th Edition.
Quantitative Business Analysis for Decision Making Simple Linear Regression.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 15-1 Chapter 15 Multiple Regression Model Building Basic Business Statistics 11 th Edition.
Regression Model Building Setting: Possibly a large set of predictor variables (including interactions). Goal: Fit a parsimonious model that explains variation.
Generalized Linear Models
Multiple Regression continued… STAT E-150 Statistical Methods.
Copyright ©2011 Pearson Education 15-1 Chapter 15 Multiple Regression Model Building Statistics for Managers using Microsoft Excel 6 th Global Edition.
STAT E-150 Statistical Methods
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Slides by JOHN LOUCKS & Updated by SPIROS VELIANITIS.
Inference for regression - Simple linear regression
Chapter 13: Inference in Regression
Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall 15-1 Chapter 15 Multiple Regression Model Building Statistics for Managers using Microsoft.
Chapter 12 Multiple Regression and Model Building.
Copyright © 2013, 2010 and 2007 Pearson Education, Inc. Chapter Inference on the Least-Squares Regression Model and Multiple Regression 14.
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
1 1 Slide Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple Coefficient of Determination n Model Assumptions n Testing.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
Business Statistics for Managerial Decision Farideh Dehkordi-Vakil.
AN INTRODUCTION TO LOGISTIC REGRESSION ENI SUMARMININGSIH, SSI, MM PROGRAM STUDI STATISTIKA JURUSAN MATEMATIKA UNIVERSITAS BRAWIJAYA.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
April 4 Logistic Regression –Lee Chapter 9 –Cody and Smith 9:F.
Week 5: Logistic regression analysis Overview Questions from last week What is logistic regression analysis? The mathematical model Interpreting the β.
1 1 Slide © 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
1 11 Simple Linear Regression and Correlation 11-1 Empirical Models 11-2 Simple Linear Regression 11-3 Properties of the Least Squares Estimators 11-4.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 13 Multiple Regression Section 13.3 Using Multiple Regression to Make Inferences.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Model Building and Model Diagnostics Chapter 15.
1 1 Slide © 2011 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole.
Chapter 14: Inference for Regression. A brief review of chapter 4... (Regression Analysis: Exploring Association BetweenVariables )  Bi-variate data.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
Logistic Regression. Linear regression – numerical response Logistic regression – binary categorical response eg. has the disease, or unaffected by the.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 14-1 Chapter 14 Introduction to Multiple Regression Basic Business Statistics 10 th Edition.
Logistic Regression Analysis Gerrit Rooks
Dates Presentations Wed / Fri Ex. 4, logistic regression, Monday Dec 7 th Final Tues. Dec 8 th, 3:30.
Logistic Regression Saed Sayad 1www.ismartsoft.com.
Introduction to Multiple Regression Lecture 11. The Multiple Regression Model Idea: Examine the linear relationship between 1 dependent (Y) & 2 or more.
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 14-1 Chapter 14 Multiple Regression Model Building Statistics for Managers.
Logistic Regression and Odds Ratios Psych DeShon.
Nonparametric Statistics
Logistic Regression Logistic Regression - Binary Response variable and numeric and/or categorical explanatory variable(s) –Goal: Model the probability.
LOGISTIC REGRESSION. Purpose  Logistical regression is regularly used when there are only two categories of the dependent variable and there is a mixture.
Logistic Regression: Regression with a Binary Dependent Variable.
Stats Methods at IC Lecture 3: Regression.
BINARY LOGISTIC REGRESSION
Logistic Regression When and why do we use logistic regression?
Chapter 14 Inference on the Least-Squares Regression Model and Multiple Regression.
Regression Analysis AGEC 784.
Logistic Regression APKC – STATS AFAC (2016).
John Loucks St. Edward’s University . SLIDES . BY.
Generalized Linear Models
Chapter 6 Logistic Regression: Regression with a Binary Dependent Variable Copyright © 2010 Pearson Education, Inc., publishing as Prentice-Hall.
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

Multiple Logistic Regression STAT E-150 Statistical Methods

2 Multiple Logistic Regression is used when there are several predictors, and the response variable is a binary variable. As we did previously, we can use an indicator variable to represent the response variable, with 1 to represent the presence of some condition (“success” or "yes") and 0 to represent the absence of the condition (“failure” or "no"): The logistic regression model describes how the probability of “success” is related to the values of the explanatory variables, which can be categorical or quantitative.

3 The Multiple Logistic Regression Model Logit form: Probability form:

4 Conditions: Linearity: Check for a linear relationship between a predictor and the logit - transformed response variable, using logit plots. Probability Model: The response values must be random and independent. Think carefully about how the data was produced.

5 As with other regression methods, if the conditions are satisfied, we can test hypotheses and construct confidence intervals, and use the results to describe relationships and make predictions. There are advantages to this analysis. First, there are no assumptions about the distributions of the predictors; they do not have to be normally distributed, linearly related, or have equal variances within each group. In addition, there are no restrictions on the type of predictors.

6 Let's return to the example we discussed earlier: Suppose that the sales director of appliance stores wants to find out which factors encourage customers to purchase extended warranties after a major appliance purchase. The response variable indicates whether a warranty is purchased. The predictor variables are - Customer gender - Age of the customer - Whether a gift is offered with the warranty - Price of the appliance - Race of the customer (this is coded with four indicator variables to represent White, African-American, Hispanic, and Other)

Variables in the Equation BS.E.WalddfSig.Exp(B) Step 1 a Gender Gift Age Price White AfricanAmerican Hispanic Constant a. Variable(s) entered on step 1: Gender, Gift, Age, Price, White, AfricanAmerican, Hispanic. 7 Let‘s start with the full model, using all predictors: The significance of each predictor is measured using the Wald statistic. (Note that SPSS finds a Wald Chi-Square and not a Wald z which you may see elsewhere; remember that the z value is just the square root of the Chi-Square value. In this case, use the sign of the corresponding coefficient estimate, β i.)

Variables in the Equation BS.E.WalddfSig.Exp(B) Step 1 a Gender Gift Age Price White AfricanAmerican Hispanic Constant a. Variable(s) entered on step 1: Gender, Gift, Age, Price, White, AfricanAmerican, Hispanic. 8 Let‘s start with the full model, using all predictors: Which predictors are significant at the.10 level of significance? Gender (p =.142) and the race variables White (p =.785), AfricanAmerican (p =.933), and Hispanic (p =.652), are not significant. (Age is marginal (p =.104), and we’ll leave it in for now.)

9 We can now repeat the analysis using only the significant predictors. Which predictors are significant in this reduced model?

10 We can now repeat the analysis using only the significant predictors. Which predictors are significant in this reduced model? The reduced model indicates that all variables are significant, even at the.05 level of significance.

11 As we have discussed, there are seemingly contradictory values: the coefficient for Price is too small to fit into three decimal places, but we know it is an important predictor because the odds ratio is 1. We can try to find more information by dividing the values of Price by 100. Here are the results for the new model:

12 Here are the results for the new model: Note that the odds ratio for Price100 is now and the new coefficient is.040. All other values are unchanged. Variables in the Equation BS.E.WalddfSig.Exp(B) Step 1 a Gift Age Price Constant a. Variable(s) entered on step 1: Gift, Age, Price100.

13 How do we assess this model? Use the -2(log likelihood) value; lower values indicate a better fit. When you are comparing models, consider the difference in the -2LL values to see if the difference is significant.

14 What if we remove the Age variable? Variables in the Equation BS.E.WalddfSig.Exp(B) Step 1 a Gift Age Price Constant a. Variable(s) entered on step 1: Gift, Age, Price100.

15 What if we remove the Age variable?

16 With the Age variable: Without the Age variable: We can conclude that the larger model is better.

17 We can also compare the results of the Hosmer-Lemenshow test for goodness-of-fit; it assesses the general model, not the model parameters. With the Age variable: Without the Age variable: Smaller p-values indicate a lack of fit for the model. Which model appears to be better based on this value?

18 We can also compare the results of the Hosmer-Lemenshow test for goodness-of-fit; it assesses the general model, not the model parameters. With the Age variable: Without the Age variable: Again we can conclude that the larger model is a better fit.