Education 795 Class Notes Applied Research Logistic Regression Note set 10.

Slides:



Advertisements
Similar presentations
Continued Psy 524 Ainsworth
Advertisements

Linear Regression.
Brief introduction on Logistic Regression
Logistic Regression Psy 524 Ainsworth.
Logistic Regression I Outline Introduction to maximum likelihood estimation (MLE) Introduction to Generalized Linear Models The simplest logistic regression.
Binary Logistic Regression: One Dichotomous Independent Variable
Logistic Regression.
Logit & Probit Regression
Logistic Regression STA302 F 2014 See last slide for copyright information 1.
Models with Discrete Dependent Variables
April 25 Exam April 27 (bring calculator with exp) Cox-Regression
Logistic Regression Multivariate Analysis. What is a log and an exponent? Log is the power to which a base of 10 must be raised to produce a given number.
Chapter 8 Logistic Regression 1. Introduction Logistic regression extends the ideas of linear regression to the situation where the dependent variable,
Introduction to Logistic Regression. Simple linear regression Table 1 Age and systolic blood pressure (SBP) among 33 adult women.
FIN357 Li1 Binary Dependent Variables Chapter 12 P(y = 1|x) = G(  0 + x  )
EPI 809/Spring Multiple Logistic Regression.
(Correlation and) (Multiple) Regression Friday 5 th March (and Logistic Regression too!)
Notes on Logistic Regression STAT 4330/8330. Introduction Previously, you learned about odds ratios (OR’s). We now transition and begin discussion of.
An Introduction to Logistic Regression
Generalized Linear Models
Regression and Correlation
Marshall University School of Medicine Department of Biochemistry and Microbiology BMS 617 Lecture 12: Multiple and Logistic Regression Marshall University.
Unit 4b: Fitting the Logistic Model to Data © Andrew Ho, Harvard Graduate School of EducationUnit 4b – Slide 1
MODELS OF QUALITATIVE CHOICE by Bambang Juanda.  Models in which the dependent variable involves two ore more qualitative choices.  Valuable for the.
Logistic Regression. Outline Review of simple and multiple regressionReview of simple and multiple regression Simple Logistic RegressionSimple Logistic.
Wednesday PM  Presentation of AM results  Multiple linear regression Simultaneous Simultaneous Stepwise Stepwise Hierarchical Hierarchical  Logistic.
Excepted from HSRP 734: Advanced Statistical Methods June 5, 2008.
Logistic Regression STA2101/442 F 2014 See last slide for copyright information.
April 6 Logistic Regression –Estimating probability based on logistic model –Testing differences among multiple groups –Assumptions for model.
Logistic Regression Database Marketing Instructor: N. Kumar.
University of Warwick, Department of Sociology, 2014/15 SO 201: SSAASS (Surveys and Statistics) (Richard Lampard) Week 7 Logistic Regression I.
CS 478 – Tools for Machine Learning and Data Mining Linear and Logistic Regression (Adapted from various sources) (e.g., Luiz Pessoa PY 206 class at Brown.
AN INTRODUCTION TO LOGISTIC REGRESSION ENI SUMARMININGSIH, SSI, MM PROGRAM STUDI STATISTIKA JURUSAN MATEMATIKA UNIVERSITAS BRAWIJAYA.
LOGISTIC REGRESSION A statistical procedure to relate the probability of an event to explanatory variables Used in epidemiology to describe and evaluate.
Logistic (regression) single and multiple. Overview  Defined: A model for predicting one variable from other variable(s).  Variables:IV(s) is continuous/categorical,
When and why to use Logistic Regression?  The response variable has to be binary or ordinal.  Predictors can be continuous, discrete, or combinations.
Linear vs. Logistic Regression Log has a slightly better ability to represent the data Dichotomous Prefer Don’t Prefer Linear vs. Logistic Regression.
April 4 Logistic Regression –Lee Chapter 9 –Cody and Smith 9:F.
Logistic Regression. Linear Regression Purchases vs. Income.
Multiple Logistic Regression STAT E-150 Statistical Methods.
Multiple Regression  Similar to simple regression, but with more than one independent variable R 2 has same interpretation R 2 has same interpretation.
LOGISTIC REGRESSION Binary dependent variable (pass-fail) Odds ratio: p/(1-p) eg. 1/9 means 1 time in 10 pass, 9 times fail Log-odds ratio: y = ln[p/(1-p)]
Logistic Regression Analysis Gerrit Rooks
Dates Presentations Wed / Fri Ex. 4, logistic regression, Monday Dec 7 th Final Tues. Dec 8 th, 3:30.
1 Introduction to Modeling Beyond the Basics (Chapter 7)
Logistic Regression An Introduction. Uses Designed for survival analysis- binary response For predicting a chance, probability, proportion or percentage.
Roger B. Hammer Assistant Professor Department of Sociology Oregon State University Conducting Social Research Logistic Regression Categorical Data Analysis.
Logistic Regression and Odds Ratios Psych DeShon.
Nonparametric Statistics
Logistic Regression For a binary response variable: 1=Yes, 0=No This slide show is a free open source document. See the last slide for copyright information.
Instructor: R. Makoto 1richard makoto UZ Econ313 Lecture notes.
LOGISTIC REGRESSION. Purpose  Logistical regression is regularly used when there are only two categories of the dependent variable and there is a mixture.
Logistic Regression: Regression with a Binary Dependent Variable.
Chapter 13 LOGISTIC REGRESSION. Set of independent variables Categorical outcome measure, generally dichotomous.
Nonparametric Statistics
BINARY LOGISTIC REGRESSION
Logistic Regression When and why do we use logistic regression?
Logistic Regression APKC – STATS AFAC (2016).
Logistic Regression.
Notes on Logistic Regression
Drop-in Sessions! When: Hillary Term - Week 1 Where: Q-Step Lab (TBC) Sign up with Alice Evans.
Generalized Linear Models
Multiple logistic regression
Jeffrey E. Korte, PhD BMTRY 747: Foundations of Epidemiology II
Nonparametric Statistics
Logistic Regression.
Introduction to Logistic Regression
Logistic Regression.
Regression Part II.
Presentation transcript:

Education 795 Class Notes Applied Research Logistic Regression Note set 10

Today’s Agenda Announcements (ours and yours) Q/A Applied Research Logistic regression

Pure vs. Applied Research Pure research ‘Pure research is that type of research which is directed towards increase of knowledge in science… where the primary aim is a fuller understanding of the subject under study rather than the application thereof’ (NSF, 1959) Applied research ‘Research carried out for the purpose of solving practical problems’ (Pedhazur & Pedhazur, 1991)

What do You Think? The pure researcher believes the applied scientists are not creative, that applied work attracts only mediocre men/women, and that applied research is like working from a cookbook. The applied researcher believes the pure scientist to be a snob, working in his/her ivory tower and afraid to put his/her findings to a real test… like Bacon’s spider, spinning webs out off his substance. (Storer, 1966, p. 108)

Sociobehavioral Research and Policy Advocacy Another endless debate over whether scientists should limit the presentation of their findings or also act as advocates for policies they presumably support…

Poem for the Day Thou shalt not answer questionnaires Or quizzes upon World-Affairs, Nor with compliance Take any test. Thou shalt not sit With statisticians nor commit A social science (Auden, 1950, p. 69)

We Turn Now to Logistic Regression When / why do we use logistic regression? Theory behind logistic regression Running logistic regression on SPSS Interpreting logistic regression analysis

When and Why? To test predictors when the outcome variable is a categorical dichotomy (yes/no, pass/fail, survive/die) Used because having a categorical dichotomy as an outcome variable violates the assumption of linearity and homogeneity in normal regression

Dichotomous Outcomes Aldrich & Nelson present two solutions to the violation of the linear regression assumptions for dichotomous outcomes Linear Probability Models Weighted Least Squares (which we will not cover) Nonlinear Probability Models Logit (Most commonly used) Probit (which we will not cover)

Nonlinear Probability Model Log (P/(1-P)=b 0 +b 1 X 1 +…+b n X n Let’s look at the left hand side P=probability of success (defined by the researcher, e.g. pass, graduate, survive) 1-P=probability of failure (1-probability of success) This ratio demands that the estimated coefficients remain positive Taking the Logarithm of this ratio restricts the range to be from 0 to 1 The left hand side is commonly referred to as the “logit” or the log(odds)

Odds Ratio P/(1-P) is called an odds. Simple Example: Hat with 5 red chips and 10 green chips. You win if you pull a red chip. Probability of winning is 1/3 Probability of not winning is 2/3 Odds of winning = 1/3 / 2/3 = ½ or 1:2 In other words, your odds of winning are 1 to 2 (there are 2 green chips for every 1 red chip, in lay terms, you are less likely to win then to lose)

Odds Ratio Let’s reverse the example Simple Example: Hat with 10 red chips and 5 green chips. You win if you pull a red chip. Probability of winning is 2/3 Probability of not winning is 1/3 Odds of winning = 2/3 / 1/3 = 2 or 2:1 In other words, your odds of winning are 2 to 1 (there are 1 green chips for every 2 red chip, in lay terms, you are more likely to win then to lose)

Morale of the Story For Odds Ratios Less than 1, success is less likely For Odds Ratios Greater than 1, success is more likely

Logistic Function Continuous Smooth S-shaped curve Takes on values between 0<=p<=1 Increases monotonically Symmetric around 0 Let’s take a look

The S-Shaped Curve P(Y=1)

Solving for P Single predictor Multiple predictors

Assumptions of Nonlinear Model Random sample Independent observations X’s are independent (minimal collinearity among the predictors) Estimation Technique called maximum likelihood estimation is used in logit models

Interpretation Issues Unlike the coefficients in a linear regression model, logistic regression results cannot be interpreted as the rate of change in the expected value the dependent variable, but the change in the probability of Y = 1 for any particular X The rate of change in the probability of Y = 1 is dependent upon the value of X (and all other Xs, if there are any in the analysis)

Interpreting Coefficients The output from SPSS will include the  ’s (the estimated log odds) and the exp (  ) the odds. How do we interpret the odds? If exp (  ) =1.6, then a one unit increase in X predicts a 60% increase in the odds of success. If exp (  ) =2.6 then a one unit increase in X predicts a 160% increase in the odds of success If exp (  ) =.7 then a one unit increase in X predicts a 30% decrease in the odds of success

A Simple Example Many natural applications of logistic regression are from medicine Understanding Coronary Heart Disease (CHD): 100 Cases 2 Variables Age (measured in years) Evidence of CHD (1 = Yes, 0 = No)

Sample Output

Sample Output 2

Output 3 Wald is similar to t-statistic Tests the null  = 0 Exp(B) indicates the change in odds resulting from a unit change in the predictor: Exp(B) > 1: as X  Probability  Exp(B) < 1: as X  Probability 

In Class Project Run a logistic regression with one predictor. Interpret the odds ratio

For Next Week Read and Discuss The Status of Women and Minorities Among Community College Faculty. Perna. L.W. (2003) Research in Higher Education 44(2) p