Linear vs. Logistic Regression Log has a slightly better ability to represent the data Dichotomous Prefer Don’t Prefer Linear vs. Logistic Regression.

Slides:



Advertisements
Similar presentations
Dummy Dependent variable Models
Advertisements

Statistical Analysis SC504/HS927 Spring Term 2008
Sociology 680 Multivariate Analysis Logistic Regression.
Lecture 11 (Chapter 9).
Linear Regression.
Brief introduction on Logistic Regression
- Word counts - Speech error counts - Metaphor counts - Active construction counts Moving further Categorical count data.
Logistic Regression Psy 524 Ainsworth.
Logistic Regression I Outline Introduction to maximum likelihood estimation (MLE) Introduction to Generalized Linear Models The simplest logistic regression.
Binary Logistic Regression: One Dichotomous Independent Variable
Logistic Regression.
Limited Dependent Variables
Regression Analysis Module 3. Regression Regression is the attempt to explain the variation in a dependent variable using the variation in independent.
Logistic Regression STA302 F 2014 See last slide for copyright information 1.
Logistic Regression Multivariate Analysis. What is a log and an exponent? Log is the power to which a base of 10 must be raised to produce a given number.
Introduction to Logistic Regression. Simple linear regression Table 1 Age and systolic blood pressure (SBP) among 33 adult women.
An Introduction to Logistic Regression JohnWhitehead Department of Economics Appalachian State University.
Notes on Logistic Regression STAT 4330/8330. Introduction Previously, you learned about odds ratios (OR’s). We now transition and begin discussion of.
An Introduction to Logistic Regression
Logistic Regression II Simple 2x2 Table (courtesy Hosmer and Lemeshow) Exposure=1Exposure=0 Disease = 1 Disease = 0.
Wednesday PM  Presentation of AM results  Multiple linear regression Simultaneous Simultaneous Stepwise Stepwise Hierarchical Hierarchical  Logistic.
Logistic Regression III: Advanced topics Conditional Logistic Regression for Matched Data Conditional Logistic Regression for Matched Data.
1 G Lect 11W Logistic Regression Review Maximum Likelihood Estimates Probit Regression and Example Model Fit G Multiple Regression Week 11.
Logistic Regression STA2101/442 F 2014 See last slide for copyright information.
Copyright © 2010 Pearson Education, Inc Chapter Seventeen Correlation and Regression.
Logistic Regression Database Marketing Instructor: N. Kumar.
Logistic regression A quick intro. Why Logistic Regression? Big idea: dependent variable is a dichotomy (though can use for more than 2 categories i.e.
1 היחידה לייעוץ סטטיסטי אוניברסיטת חיפה פרופ’ בנימין רייזר פרופ’ דוד פרג’י גב’ אפרת ישכיל.
LOGISTIC REGRESSION A statistical procedure to relate the probability of an event to explanatory variables Used in epidemiology to describe and evaluate.
When and why to use Logistic Regression?  The response variable has to be binary or ordinal.  Predictors can be continuous, discrete, or combinations.
April 4 Logistic Regression –Lee Chapter 9 –Cody and Smith 9:F.
Week 5: Logistic regression analysis Overview Questions from last week What is logistic regression analysis? The mathematical model Interpreting the β.
Journal: 1)Suppose you guessed on a multiple choice question (4 answers). What was the chance that you marked the correct answer? Explain. 2)What is the.
Logistic Regression. Linear Regression Purchases vs. Income.
Applied Epidemiologic Analysis - P8400 Fall 2002 Labs 6 & 7 Case-Control Analysis ----Logistic Regression Henian Chen, M.D., Ph.D.
LOGISTIC REGRESSION Binary dependent variable (pass-fail) Odds ratio: p/(1-p) eg. 1/9 means 1 time in 10 pass, 9 times fail Log-odds ratio: y = ln[p/(1-p)]
Logistic Regression Analysis Gerrit Rooks
Dates Presentations Wed / Fri Ex. 4, logistic regression, Monday Dec 7 th Final Tues. Dec 8 th, 3:30.
Logistic Regression Saed Sayad 1www.ismartsoft.com.
Multiple Regression Analysis Regression analysis with two or more independent variables. Leads to an improvement.
Applied Epidemiologic Analysis - P8400 Fall 2002 Labs 6 & 7 Case-Control Analysis ----Logistic Regression Henian Chen, M.D., Ph.D.
LECTURE 06: CLASSIFICATION PT. 2 February 10, 2016 SDS 293 Machine Learning.
Logistic Regression Hal Whitehead BIOL4062/5062.
Logistic Regression An Introduction. Uses Designed for survival analysis- binary response For predicting a chance, probability, proportion or percentage.
Roger B. Hammer Assistant Professor Department of Sociology Oregon State University Conducting Social Research Logistic Regression Categorical Data Analysis.
Logistic Regression and Odds Ratios Psych DeShon.
Nonparametric Statistics
Birthweight (gms) BPDNProp Total BPD (Bronchopulmonary Dysplasia) by birth weight Proportion.
Logistic Regression For a binary response variable: 1=Yes, 0=No This slide show is a free open source document. See the last slide for copyright information.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
The Probit Model Alexander Spermann University of Freiburg SS 2008.
LOGISTIC REGRESSION. Purpose  Logistical regression is regularly used when there are only two categories of the dependent variable and there is a mixture.
Logistic Regression: Regression with a Binary Dependent Variable.
Chapter 13 LOGISTIC REGRESSION. Set of independent variables Categorical outcome measure, generally dichotomous.
The Probit Model Alexander Spermann University of Freiburg SoSe 2009
Nonparametric Statistics
BINARY LOGISTIC REGRESSION
Logistic Regression APKC – STATS AFAC (2016).
Logistic Regression.
Notes on Logistic Regression
Logistic Regression Part One
Multiple logistic regression
Jeffrey E. Korte, PhD BMTRY 747: Foundations of Epidemiology II
Nonparametric Statistics
Introduction to Logistic Regression
Modeling with Dichotomous Dependent Variables
Chapter 6 Logistic Regression: Regression with a Binary Dependent Variable Copyright © 2010 Pearson Education, Inc., publishing as Prentice-Hall.
Logistic Regression.
Presentation transcript:

Linear vs. Logistic Regression Log has a slightly better ability to represent the data Dichotomous Prefer Don’t Prefer Linear vs. Logistic Regression

History: Logistic Regression Exponential

Odd’s Ratio Using the values 0 and 1 is helpful for several reasons. Among those reasons is that the values can be interpreted as probabilities

Product of All Probabilities Is Likelihood Heads or Tails: 50% Heads twice in a row: 50% * 50% = 25% The likelihood of purchases by 36 prospects is the product of the probability: P1 will buy * P2 *... * P36

The original classification table is put in here to get the Ns as well as to get the original percent among the respondents The original percent is turned into a probability The Average Odds is then multiplied by the Exp of the Beta. Which is then turned back into a percentage The original percentage is subtracted from the predicted percent to determine the change Mathematics `~1:1 Ratio for getting a No or Yes ` Logit Model Includes Log; So Need to Convert to Odds 2.52 vs Delta from the Average Odds 72% / 28% = %-72% = 28% The Regression Beta is then converted to Odds. FormulasOutputProcessCalculations ln Odds = P / (1-P) Odds – (Odds*P) = P Odds = P + Odds*P Odds = P(1 + Odds) P = Odds / (1 + Odds)

Logistic: Maximum Likelihood Logistic regression tries out different values for the intercept and the coefficients until it finds the value that results in probabilities—that is, likelihoods —that are closest to the actual, observed probabilities.

Purchase Dataset Conceptually, if a person has greater income, the probability that he or she will purchase is greater than if the person has less income.

Probability of No Purchase: Person who did not purchase has a 0 on the Purchased variable Predicted probability of 2% that he will purchase Probability of Purchase: Person who did purchase has a 1 on the Purchased variable Predicted probability of 94% that this person will purchase The probabilities are of two different events: No Purchase and Purchase In the first case, it’s 98% that he doesn’t purchase, and he doesn’t.

Measure of Goodness R^2 ranges from 0 to 1.0, and can be considered as a percentage of variability. An R 2 of 1.0—or 100%—means that 100% of the variance in the dependent variable can be explained by variability in the independent variable or variables. We use the log likelihood as our criterion for the “best” coefficients. The closer to 0.0 a log likelihood: the better the fit the closer you’ve come to maximizing the estimate of the likelihood.