 Seeks to determine group membership from predictor variables ◦ Given group membership, how many people can we correctly classify?

Slides:



Advertisements
Similar presentations
Items to consider - 3 Multicollinearity
Advertisements

Logistic Regression Psy 524 Ainsworth.
Chapter 17 Making Sense of Advanced Statistical Procedures in Research Articles.
CJT 765: Structural Equation Modeling Class 3: Data Screening: Fixing Distributional Problems, Missing Data, Measurement.
Chapter 17 Overview of Multivariate Analysis Methods
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 14 Using Multivariate Design and Analysis.
Linear Discriminant Function LDF & MANOVA LDF & Multiple Regression Geometric example of LDF & multivariate power Evaluating & reporting LDF results 3.
Discrim Continued Psy 524 Andrew Ainsworth. Types of Discriminant Function Analysis They are the same as the types of multiple regression Direct Discrim.
Multivariate Data Analysis Chapter 4 – Multiple Regression.
ANCOVA Psy 420 Andrew Ainsworth. What is ANCOVA?
Lecture 6: Multiple Regression
MANOVA LDF & MANOVA Geometric example of MANVOA & multivariate power MANOVA dimensionality Follow-up analyses if k > 2 Factorial MANOVA.
Discriminant Analysis – Basic Relationships
Multiple-group linear discriminant function maximum & contributing ldf dimensions concentrated & diffuse ldf structures follow-up analyses evaluating &
What Is Multivariate Analysis of Variance (MANOVA)?
An Introduction to Logistic Regression
LECTURE 17 MANOVA. Other Measures Pillai-Bartlett trace, V Multiple discriminant analysis (MDA) is the part of MANOVA where canonical roots are.
Discriminant Analysis Testing latent variables as predictors of groups.
SW388R7 Data Analysis & Computers II Slide 1 Multiple Regression – Basic Relationships Purpose of multiple regression Different types of multiple regression.
Discriminant analysis
SW388R7 Data Analysis & Computers II Slide 1 Discriminant Analysis – Basic Relationships Discriminant Functions and Scores Describing Relationships Classification.
Quantitative Business Analysis for Decision Making Multiple Linear RegressionAnalysis.
ANCOVA Lecture 9 Andrew Ainsworth. What is ANCOVA?
Marketing Research Aaker, Kumar, Day and Leone Tenth Edition
Selecting the Correct Statistical Test
Discriminant Function Analysis Basics Psy524 Andrew Ainsworth.
بسم الله الرحمن الرحیم.. Multivariate Analysis of Variance.
Chapter Eighteen Discriminant Analysis Chapter Outline 1) Overview 2) Basic Concept 3) Relation to Regression and ANOVA 4) Discriminant Analysis.
Chapter 18 Four Multivariate Techniques Angela Gillis & Winston Jackson Nursing Research: Methods & Interpretation.
Discriminant Analysis
Multiple Regression The Basics. Multiple Regression (MR) Predicting one DV from a set of predictors, the DV should be interval/ratio or at least assumed.
11/12/2012ISC471 / HCI571 Isabelle Bichindaritz 1 Prediction.
Regression Analyses. Multiple IVs Single DV (continuous) Generalization of simple linear regression Y’ = b 0 + b 1 X 1 + b 2 X 2 + b 3 X 3...b k X k Where.
Discriminant Analysis Discriminant analysis is a technique for analyzing data when the criterion or dependent variable is categorical and the predictor.
17-1 COMPLETE BUSINESS STATISTICS by AMIR D. ACZEL & JAYAVEL SOUNDERPANDIAN 6 th edition (SIE)
Statistical Analysis of Data1 of 38 1 of 42 Department of Cognitive Science Adv. Experimental Methods & Statistics PSYC 4310 / COGS 6310 MANOVA Multivariate.
Adjusted from slides attributed to Andrew Ainsworth
Model Selection and Validation. Model-Building Process 1. Data collection and preparation 2. Reduction of explanatory or predictor variables (for exploratory.
Multivariate Data Analysis Chapter 1 - Introduction.
Review for Final Examination COMM 550X, May 12, 11 am- 1pm Final Examination.
» So, I’ve got all this data…what now? » Data screening – important to check for errors, assumptions, and outliers. » What’s the most important? ˃Depends.
Assumptions 5.4 Data Screening. Assumptions Parametric tests based on the normal distribution assume: – Independence – Additivity and linearity – Normality.
Descriptions. Description Correlation – simply finding the relationship between two scores ○ Both the magnitude (how strong or how big) ○ And direction.
Discriminant Analysis Dr. Satyendra Singh Professor and Director University of Winnipeg, Canada
Outline of Today’s Discussion 1.Introduction to Discriminant Analysis 2.Assumptions for Discriminant Analysis 3.Discriminant Analysis in SPSS.
Linear Regression Chapter 7. Slide 2 What is Regression? A way of predicting the value of one variable from another. – It is a hypothetical model of the.
Two-Group Discriminant Function Analysis. Overview You wish to predict group membership. There are only two groups. Your predictor variables are continuous.
ANCOVA.
D/RS 1013 Discriminant Analysis. Discriminant Analysis Overview n multivariate extension of the one-way ANOVA n looks at differences between 2 or more.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 18 Multivariate Statistics.
MANOVA Lecture 12 Nuance stuff Psy 524 Andrew Ainsworth.
PROFILE ANALYSIS. Profile Analysis Main Point: Repeated measures multivariate analysis One/Several DVs all measured on the same scale.
DISCRIMINANT ANALYSIS. Discriminant Analysis  Discriminant analysis builds a predictive model for group membership. The model is composed of a discriminant.
STATISTICAL TESTS USING SPSS Dimitrios Tselios/ Example tests “Discovering statistics using SPSS”, Andy Field.
Linear Discriminant Function Classification vs. Prediction Classification & ANOVA Classification Cutoffs, Errors, etc. Multivariate Classification & LDF.
Data Screening. What is it? Data screening is very important to make sure you’ve met all your assumptions, outliers, and error problems. Each type of.
Regression. Why Regression? Everything we’ve done in this class has been regression: When you have categorical IVs and continuous DVs, the ANOVA framework.
Chapter 13 LOGISTIC REGRESSION. Set of independent variables Categorical outcome measure, generally dichotomous.
Chapter 12 REGRESSION DIAGNOSTICS AND CANONICAL CORRELATION.
BINARY LOGISTIC REGRESSION
MANOVA Dig it!.
Logistic Regression APKC – STATS AFAC (2016).
Correlation, Bivariate Regression, and Multiple Regression
Multiple Regression Prof. Andy Field.
Regression.
Multivariate Analysis Lec 4
Making Sense of Advanced Statistical Procedures in Research Articles
Multiple Regression – Part II
Discriminant Analysis – Basic Relationships
Linear Discriminant Analysis
Presentation transcript:

 Seeks to determine group membership from predictor variables ◦ Given group membership, how many people can we correctly classify?

 Can think of this analysis as MANOVA turned inside out. WomenMen Left Right Right Women Right Men Left Woman Left Men

 Given your mean, can we predict what group you would be in?  IVs = predictors  DVs = groups, grouping variables

 You have classifications – 1 (groups basically) ways to linearly combine IVs to create group membership ◦ So if you have 3 groups, then there are several ways to combine variables to classify

 If you have a third group, you may have a second discriminate function.  Trying to understand what IVs are used to discriminate between groups can be kind of limited here

 Find a function to predict group membership.

 Significant predictions – can membership be predicted reliably, can we do better than chance?

 Number of significant discriminate functions ◦ Groups/classifications – 1 is the max number you will get ◦ 1 st discriminate function accounts for the most variance, provides best separation between groups ◦ 2 nd function is orthogonal, but may still be significant

 Dimensions of discrimination – what are the IVs that separate the groups? ◦ What is the pattern of relationships of IVs to the discriminate function (similar to EFA).

 Classification function – how can we weight scores to discriminate (creates regression equations)

 Adequacy of classification – how many cases are classified correctly? ◦ When there are mistakes, where do they happen?

 Effect size? ◦ Finds a canonical correlate with each function, so you can see how much of the variance in groups is accounted for by each discriminate function.

 Which predictors are the most important?  If covariates, what IVs are important after the controls?  Estimation of group means (centroids) – which means do the discriminate functions separate?

 Are the groups naturally occurring or did we randomly assign them?

 Since it’s mostly about classification, it’s ok if distributions are a bit weird as long as the discriminate function is good.  Whenever MANOVA works best, discriminate works best.

 Unequal N – not a big deal. ◦ But does influence with very small cells ( ◦ Discriminate function will be biased because that probability of that cell is so small ◦ Sample size of smallest group > IVs

 Missing data – needs to be replaced or eliminated. ◦ Dependent data ◦ Category data

 Normality ◦ Robust but we assume linear combinations of the IVs are normal – but not a good way to test this idea. ◦ Start to have problems if there are unequal N and the sample size is small.  Want 20 cases in smallest sample.

 Outliers – you are trying to predict an individuals’ group, so outliers = no good. ◦ Univariate and multivariate outliers need to eliminated. ◦ BUT run outliers separately for each classification/group.

 Homogeneity – Box’s M is still sensitive ◦ Can also check out scatter plots of scores on 1 st and 2 nd discriminate functions separated for each group (SPSS plots) ◦ If fails:  Use nonparametric – log regression

 Linearity – making linear combinations of IVs AND we are making regressions. Definitely need. ◦ Less serious errors because it just reduces power.

 Multicollinearity – since this is regression you do not want 2 predictors that measure the same thing.

 Same as MANOVA (to a point).  MANOVA – tests if there are differences in combinations of means (DVs) for groups (IVs) ◦ Then tests which DVs are best for group separation.

 Discriminate function – ◦ D = dz1 + dz2 + dz3 … ◦ D = discriminate score ◦ d = discriminate function coefficient  Get by doing canonical variates  Basically canonical correlation as: Group member predictor

 Discriminate function – ◦ D = dz1 + dz2 + dz3 … ◦ d = chosen to maximize group differences  Very similar to beta ◦ Z = DVs – they are z-scored because then that makes it easy to see ds weight in equation and gives D:  SD = 1, Mean = 0  Important for categorical prediction

 Separates as so:  0-> 1 = group 1  -1 -> 0 = group 2

 Classification equation ◦ You are assigned to the group where you have the highest classification score

 Classification ◦ C = Constant + cX1 + cX2 ◦ C = classification group ◦ c = classification coefficient ◦ X = raw score

 Standard (direct) – each predictor enters the equation at the same time and only assigned unique variance ◦ Test of means = MANOVA ◦ Test of discriminate functions – canonical correlation

 Sequential (hierarchical) – you determine order predictors enter discriminate function ◦ You are testing if a new predictor adds better classification to this equation  Similar to MANCOVA  Good for smaller number of predictors and theory driven arguments  (unfortunately, there’s not a good way to do this in SPSS, instead do it as a hierarchical regression where DV is coded as 0 and 1).

 Stepwise (statistical) – predictors enter equation based on some cut off you use. ◦ If you have 10 of the same predictors, it might be a good way to eliminate overlapping ones. ◦ But but! Dependent on sample you select ◦ You can use R2, F, change in group centers, etc.

 Inference ◦ Criteria for overall statistical significance  Use Wilk’s lambda as with MANOVA since it’s the same test.  Stepwise – you get two more options  Mahalanobis D2 and Rao’s V based on group centroid differences

 Inference ◦ Number of discriminate functions  If you have a lot of groups, you’ll get several discriminate functions but they may not be significant, usually around 2 are significant.  Evaluates like canonical correlation – eigenvalues, % of variance, chi-square

 Interpretation Discriminate 1 Discriminate 2 Dots are discriminate function centroids (means) of each group

 Cross validation – see if your discriminate function correctly classifies a new sample ◦ Split half testing ◦ Can do this by jackknifed classification (leave-it- out option in SPSS).