Ordinal Logistic Regression

Slides:



Advertisements
Similar presentations
Continued Psy 524 Ainsworth
Advertisements

Logistic Regression Psy 524 Ainsworth.
SPSS Session 5: Association between Nominal Variables Using Chi-Square Statistic.
Analysis of variance (ANOVA)-the General Linear Model (GLM)
Logit & Probit Regression
Chapter 11 Contingency Table Analysis. Nonparametric Systems Another method of examining the relationship between independent (X) and dependant (Y) variables.
Statistical Tests Karen H. Hagglund, M.S.
Logistic Regression Multivariate Analysis. What is a log and an exponent? Log is the power to which a base of 10 must be raised to produce a given number.
Some Terms Y =  o +  1 X Regression of Y on X Regress Y on X X called independent variable or predictor variable or covariate or factor Which factors.
Lecture 3: Chi-Sqaure, correlation and your dissertation proposal Non-parametric data: the Chi-Square test Statistical correlation and regression: parametric.
Statistics II: An Overview of Statistics. Outline for Statistics II Lecture: SPSS Syntax – Some examples. Normal Distribution Curve. Sampling Distribution.
CHI-SQUARE TEST OF INDEPENDENCE
Multinomial Logistic Regression
Descriptive Statistics Primer
Logistic Regression Biostatistics 510 March 15, 2007 Vanessa Perez.
Notes on Logistic Regression STAT 4330/8330. Introduction Previously, you learned about odds ratios (OR’s). We now transition and begin discussion of.
Social Science Research Design and Statistics, 2/e Alfred P. Rovai, Jason D. Baker, and Michael K. Ponton McNemar Test PowerPoint Prepared by Alfred P.
Data Analysis Statistics. Inferential statistics.
Today Concepts underlying inferential statistics
Data Analysis Statistics. Levels of Measurement Nominal – Categorical; no implied rankings among the categories. Also includes written observations and.
Summary of Quantitative Analysis Neuman and Robson Ch. 11
Multinomial Logistic Regression Basic Relationships
Review for Exam 2 Some important themes from Chapters 6-9 Chap. 6. Significance Tests Chap. 7: Comparing Two Groups Chap. 8: Contingency Tables (Categorical.
Statistical hypothesis testing – Inferential statistics II. Testing for associations.
Chapter 12 Inferential Statistics Gay, Mills, and Airasian
The Practice of Social Research
Leedy and Ormrod Ch. 11 Gray Ch. 14
Analyzing Data: Bivariate Relationships Chapter 7.
Understanding Research Results
Categorical Data Prof. Andy Field.
Hypothesis Testing Charity I. Mulig. Variable A variable is any property or quantity that can take on different values. Variables may take on discrete.
1 Psych 5500/6500 Chi-Square (Part Two) Test for Association Fall, 2008.
Copyright © 2013, 2009, and 2007, Pearson Education, Inc. Chapter 14 Comparing Groups: Analysis of Variance Methods Section 14.2 Estimating Differences.
Hierarchical Binary Logistic Regression
Understanding Statistics
9/23/2015Slide 1 Published reports of research usually contain a section which describes key characteristics of the sample included in the study. The “key”
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 13: Nominal Variables: The Chi-Square and Binomial Distributions.
بسم الله الرحمن الرحیم.. Multivariate Analysis of Variance.
Multinomial Logistic Regression Basic Relationships
Chi-square Test of Independence Steps in Testing Chi-square Test of Independence Hypotheses.
Social Science Research Design and Statistics, 2/e Alfred P. Rovai, Jason D. Baker, and Michael K. Ponton Pearson Chi-Square Contingency Table Analysis.
Multiple Regression and Model Building Chapter 15 Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
April 4 Logistic Regression –Lee Chapter 9 –Cody and Smith 9:F.
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT OSMAN BIN SAIF Session 26.
Section 9-1: Inference for Slope and Correlation Section 9-3: Confidence and Prediction Intervals Visit the Maths Study Centre.
Categorical data analysis: An overview of statistical techniques AnnMaria De Mars The Julia Group AnnMaria De Mars The Julia Group.
Elang 273: Statistics. Review: Scientific Method 1. Observe something 2. Speculated why it is so and form hypothesis 3. Test hypothesis by getting data.
12/23/2015Slide 1 The chi-square test of independence is one of the most frequently used hypothesis tests in the social sciences because it can be used.
IMPORTANCE OF STATISTICS MR.CHITHRAVEL.V ASST.PROFESSOR ACN.
Logistic Regression Analysis Gerrit Rooks
Qualitative and Limited Dependent Variable Models ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
1 Week 3 Association and correlation handout & additional course notes available at Trevor Thompson.
Jump to first page Inferring Sample Findings to the Population and Testing for Differences.
Nonparametric Statistics
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
DISCRIMINANT ANALYSIS. Discriminant Analysis  Discriminant analysis builds a predictive model for group membership. The model is composed of a discriminant.
Chapter 15 Analyzing Quantitative Data. Levels of Measurement Nominal measurement Involves assigning numbers to classify characteristics into categories.
NURS 306, Nursing Research Lisa Broughton, MSN, RN, CCRN RESEARCH STATISTICS.
Nonparametric Statistics
BINARY LOGISTIC REGRESSION
EHS Lecture 14: Linear and logistic regression, task-based assessment
Advanced Quantitative Techniques
Notes on Logistic Regression
Advanced Quantitative Techniques
Categorical Data Aims Loglinear models Categorical data
Nonparametric Statistics
Logistic Regression.
Presentation transcript:

Ordinal Logistic Regression

Read the data Describe the data use http://www.ats.ucla.edu/stat/stata/dae/ologit.dta, clear Describe the data Codebook Summarize Tabulate Make graphs

Description of the Data Dependent apply (coded 0, 1, 2) Independent pared, which is a 0/1 variable indicating whether at least one parent has a graduate degree; public, which is a 0/1 variable where 1 indicates that the undergraduate institution is a public university and 0 indicates that it is a private university gpa, which is the student's grade point average.

Using the Ordinal Logistic Model Before we run our ordinal logistic model, we will see if any cells (created by the crosstab of our categorical and response variables) are empty or extremely small.  

None of the cells is too small or empty (has no cases), so we will run our model.

Likelihood ratio chi-square (24.18) Log likelihood (-358.51244) can be used in comparisons of nested models, but we won't show an example of that here.  Observations (400) Fewer observations would have been used if any of our variables had missing values.  Likelihood ratio chi-square (24.18) our model as a whole is statistically significant, as compared to model with no predictors.  Pseudo-R-squared

Coefficients Both pared and gpa are statistically significant; public is not. 

We can obtain odds ratios using the or option after the ologit command.

For pared, we would say that for a one unit increase in pared, i. e For pared, we would say that for a one unit increase in pared, i.e., going from 0 to 1, the odds of high apply versus the combined middle and low categories are 2.85 greater, given that all of the other variables in the model are held constant.  For a one unit increase in gpa, the odds of the low and middle categories of apply versus the high category of apply are 1.85 times greater, given that the other variables in the model are held constant.  Because of the proportional odds assumption, the same increase, 1.85 times, is found between low apply and the combined categories of middle and high apply.

You can also use the listcoeff command to obtain the odds ratios, as well as the change in the odds for a standard deviation of the variable.  You can use the percent option to see the percent change in the odds.

Proportional Odds Assumption Ordinal logistic regression assumes that the coefficients that describe the relationship between, say, the lowest versus all higher categories of the response variable are the same as those that describe the relationship between the next lowest category and all higher categories, etc.  This is called the proportional odds assumption or the parallel regression assumption.  Because the relationship between all pairs of groups is the same, there is only one set of coefficients (only one model).  If this was not the case, we would need different models to describe the relationship between each pair of outcome groups. 

Test Proportional Odds Assumption We need to download a user-written command called omodel (type findit omodel).  The first test that we will show does a likelihood ratio test.  The null hypothesis is that there is no difference in the coefficients between models, so we "hope" to get a non-significant result.  The brant command performs a Brant test.  As the note at the bottom of the output indicates, we also "hope" that these tests are non-significant. 

Both of the above tests indicate that we have not violated the proportional odds assumption.  If we had, we would want to run our model as a generalized ordered logistic model using gologit2.  You need to download gologit2 by typing findit gologit2.

Predicted Probability Use pared as an example with a categorical predictor. The predicted probability of  being in the lowest category of apply is 0.59 if neither parent has a graduate level education and 0.34 otherwise.  For the middle category of apply, the predicted probabilities are 0.33 and 0.46, and for the highest category of apply, 0.079 and 0.196.  Beneath each output, we can see the values at which the variables are held; by default, they are held at their mean.

The predicted probabilities for gpa at 2, 3 and 4.  The highest predicted probability is for the lowest category of apply, which makes sense because most respondents are in that category.  You can also see that the predicted probability increases for both the middle and highest categories of apply as gpa increases. 

Below, we use the prvalue command to set the values of all of our predictor variables.  This is useful when you want to describe profiles of respondents.