1 G89.2229 Lect 11W Logistic Regression Review Maximum Likelihood Estimates Probit Regression and Example Model Fit G89.2229 Multiple Regression Week 11.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Dummy Dependent variable Models
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Continued Psy 524 Ainsworth
Linear Regression.
Brief introduction on Logistic Regression
The %LRpowerCorr10 SAS Macro Power Estimation for Logistic Regression Models with Several Predictors of Interest in the Presence of Covariates D. Keith.
Logistic Regression Psy 524 Ainsworth.
Discrete Choice Modeling William Greene Stern School of Business IFS at UCL February 11-13, 2004
Logistic Regression I Outline Introduction to maximum likelihood estimation (MLE) Introduction to Generalized Linear Models The simplest logistic regression.
Logistic Regression.
Binary Response Lecture 22 Lecture 22.
GRA 6020 Multivariate Statistics; The Linear Probability model and The Logit Model (Probit) Ulf H. Olsson Professor of Statistics.
Generalised linear models
In previous lecture, we dealt with the unboundedness problem of LPM using the logit model. In this lecture, we will consider another alternative, i.e.
1 G Lect 11M Binary outcomes in psychology Can Binary Outcomes Be Studied Using OLS Multiple Regression? Transforming the binary outcome Logistic.
An Introduction to Logistic Regression
Generalized Linear Models
Logistic regression for binary response variables.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
1 BINARY CHOICE MODELS: PROBIT ANALYSIS In the case of probit analysis, the sigmoid function is the cumulative standardized normal distribution.
HSRP 734: Advanced Statistical Methods June 19, 2008.
CS433: Modeling and Simulation Dr. Anis Koubâa Al-Imam Mohammad bin Saud University 15 October 2010 Lecture 05: Statistical Analysis Tools.
April 6 Logistic Regression –Estimating probability based on logistic model –Testing differences among multiple groups –Assumptions for model.
9-1 MGMG 522 : Session #9 Binary Regression (Ch. 13)
Generalized Linear Models All the regression models treated so far have common structure. This structure can be split up into two parts: The random part:
Linear correlation and linear regression + summary of tests
Logistic Regression. Conceptual Framework - LR Dependent variable: two categories with underlying propensity (yes/no) (absent/present) Independent variables:
When and why to use Logistic Regression?  The response variable has to be binary or ordinal.  Predictors can be continuous, discrete, or combinations.
April 4 Logistic Regression –Lee Chapter 9 –Cody and Smith 9:F.
Week 5: Logistic regression analysis Overview Questions from last week What is logistic regression analysis? The mathematical model Interpreting the β.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
1 GCRC Data Analysis with SPSS Workshop Session 5 Follow Up on FEV data Binary and Categorical Outcomes 2x2 tables 2xK tables JxK tables Logistic Regression.
Generalized Linear Models (GLMs) and Their Applications.
1 G Lect 7a G Lecture 7a Comparing proportions from independent samples Analysis of matched samples Small samples and 2  2 Tables Strength.
LOGISTIC REGRESSION Binary dependent variable (pass-fail) Odds ratio: p/(1-p) eg. 1/9 means 1 time in 10 pass, 9 times fail Log-odds ratio: y = ln[p/(1-p)]
Logistic Regression Analysis Gerrit Rooks
Logistic regression (when you have a binary response variable)
ALISON BOWLING MAXIMUM LIKELIHOOD. GENERAL LINEAR MODEL.
Roger B. Hammer Assistant Professor Department of Sociology Oregon State University Conducting Social Research Logistic Regression Categorical Data Analysis.
Nonparametric Statistics
The Probit Model Alexander Spermann University of Freiburg SS 2008.
Construction Engineering 221 Probability and Statistics.
R Programming/ Binomial Models Shinichiro Suna. Binomial Models In binomial model, we have one outcome which is binary and a set of explanatory variables.
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Instructor: R. Makoto 1richard makoto UZ Econ313 Lecture notes.
LOGISTIC REGRESSION. Purpose  Logistical regression is regularly used when there are only two categories of the dependent variable and there is a mixture.
Chapter 13 LOGISTIC REGRESSION. Set of independent variables Categorical outcome measure, generally dichotomous.
The Probit Model Alexander Spermann University of Freiburg SoSe 2009
Sampling and Sampling Distributions
Chapter 4: Basic Estimation Techniques
BINARY LOGISTIC REGRESSION
Statistical Modelling
Logistic Regression When and why do we use logistic regression?
Logistic Regression APKC – STATS AFAC (2016).
CHAPTER 7 Linear Correlation & Regression Methods
Logistic Regression.
Notes on Logistic Regression
Basic Estimation Techniques
THE LOGIT AND PROBIT MODELS
Generalized Linear Models
Multiple logistic regression
THE LOGIT AND PROBIT MODELS
Nonparametric Statistics
ביצוע רגרסיה לוגיסטית. פרק ה-2
Introduction to Logistic Regression
Modeling with Dichotomous Dependent Variables
Logistic Regression.
Presentation transcript:

1 G Lect 11W Logistic Regression Review Maximum Likelihood Estimates Probit Regression and Example Model Fit G Multiple Regression Week 11 (Wednesday)

2 G Lect 11W Logistic Regression Review Additive probability models are often not ideal. »p is bounded by [0,1]. Log odds of p [  (p)] has nice properties. »Unbounded, with  (p)=0 for p=.5. Important equations

3 G Lect 11W Logistic Example: Clinic Only

4 G Lect 11W Maximum Likelihood Estimates Strategy: Choose guesses for regression weights that makes the outcome seem likely under the model. We use simple binomial probability theory to calculate the likelihood. »Choose some probability starting values »Calculate how likely each observation is under the initial probabilities »Summarize the overall likelihood »Modify the model to improve the likelihood

5 G Lect 11W What is the likelihood of rolling a "1" on a single die? What is the likelihood of rolling "snake eyes" with two dice? Review of Binomial Probability 1/6

6 G Lect 11W What is the likelihood of rolling one "1" and one value other than "1"? What is the probability of Y (where Y is 1 when die is "1" and 0 otherwise)? Review of Binomial Probability

7 G Lect 11W Suppose Suppose Y is a binary random variable with probability p. What is the probability of Y? Suppose that r=  Y, where n observations are summed. What is probability of  Y =r? Review of Binomial Probability

8 G Lect 11W Summarizing Likelihood Probabilities of rare events get small. The probability that a particular sample will have a specific pattern of values is very small. Products of probabilities get exponentially smaller Keeping track of these small numbers is facilitated by taking logs. Ln likelihoods sum (whereas likelihoods need to be multiplied)

9 G Lect 11W Example This Excel spreadsheet shows how the regression coefficients change the fitted probabilities.

10 G Lect 11W Probit as an Alternative to Logit Instead of using p=(1+e -  ) -1, some statisticians have suggested using the cumulative normal function p=  (z). »z = 0 for p=.5 »z=> -∞ for p=>0 »z=> ∞ for p=>1 The linear model is written for the assumed standard normal variable underlying the binary process. »The process is made binary by applying some discrete threshold.

11 G Lect 11W Probit Example: Clinic Only * * * * * * * * * * * * P R O B I T A N A L Y S I S * * * * * Parameter estimates converged after 13 iterations. Optimal solution found. Parameter Estimates (PROBIT model: (PROBIT(p)) = Intercept + BX): Regression Coeff. Standard Error Coeff./S.E. FEMALE AGEC Intercept Standard Error Intercept/S.E

12 G Lect 11W The relation of Probit to Logit

13 G Lect 11W Fit Statistics in Logistic Regression Likelihood ratio test of k predictors »Compare LL(model) to LL(null) LL(null) has only intercept in model -2*[LL(model-LL(null)] is distributed as Chi Square on k df Quasi R-square

14 G Lect 11W Other Fit Measures Cox & Snell Index Nagelkerke Index »Scales R 2 CS to have maximum of 1.0 Akaike's Information Criterion

15 G Lect 11W Hosmer-Lemeshow test Hosmer and Lemeshow argue that if the linear logistic model is correct, then we should be able to »Break the sample into g groups »Compute the average probability of the outcome in each group »Compare the observed frequencies to the expected using a Pearson Chi Square »Evaluate the Chi Square statistic to a central Chi Square on g-2 df. Test is useful when sample size is large.