Introduction to Multivariate Analysis of Variance, Factor Analysis, and Logistic Regression Rubab G. ARIM, MA University of British Columbia December 2006.

Slides:



Advertisements
Similar presentations
Multiple Analysis of Variance – MANOVA
Advertisements

Analysis of variance (ANOVA)-the General Linear Model (GLM)
Chapter Nineteen Factor Analysis.
CPSY 501: Lec13, 28Nov MANOVA: reading and interpreting Lecture notes from Rubab Arim (UBC) Sample journal article: Lecture notes from Jess Nee Range,
© LOUIS COHEN, LAWRENCE MANION & KEITH MORRISON
SPSS Series 3: Repeated Measures ANOVA and MANOVA
Lecture 7: Principal component analysis (PCA)
Psychology 202b Advanced Psychological Statistics, II April 7, 2011.
WENDIANN SETHI SPRING 2011 SPSS ADVANCED ANALYSIS.
Analysis of variance (ANOVA)-the General Linear Model (GLM)
Factor Analysis There are two main types of factor analysis:
Discrim Continued Psy 524 Andrew Ainsworth. Types of Discriminant Function Analysis They are the same as the types of multiple regression Direct Discrim.
Multiple Regression Models Advantages of multiple regression Important preliminary analyses Parts of a multiple regression model & interpretation Differences.
19-1 Chapter Nineteen MULTIVARIATE ANALYSIS: An Overview.
“Ghost Chasing”: Demystifying Latent Variables and SEM
Education 795 Class Notes Factor Analysis II Note set 7.
Summary of Quantitative Analysis Neuman and Robson Ch. 11
Multivariate Methods EPSY 5245 Michael C. Rodriguez.
Leedy and Ormrod Ch. 11 Gray Ch. 14
Factor Analysis Psy 524 Ainsworth.
Example of Simple and Multiple Regression
Social Science Research Design and Statistics, 2/e Alfred P. Rovai, Jason D. Baker, and Michael K. Ponton Factor Analysis PowerPoint Prepared by Alfred.
SPSS Series 1: ANOVA and Factorial ANOVA
By Hui Bian Office for Faculty Excellence 1. K-group between-subjects MANOVA with SPSS Factorial between-subjects MANOVA with SPSS How to interpret SPSS.
بسم الله الرحمن الرحیم.. Multivariate Analysis of Variance.
Chapter 12 Examining Relationships in Quantitative Research Copyright © 2013 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin.
Soc 3306a Lecture 9: Multivariate 2 More on Multiple Regression: Building a Model and Interpreting Coefficients.
Advanced Correlational Analyses D/RS 1013 Factor Analysis.
Factor Analysis Psy 524 Ainsworth. Assumptions Assumes reliable correlations Highly affected by missing data, outlying cases and truncated data Data screening.
Thursday AM  Presentation of yesterday’s results  Factor analysis  A conceptual introduction to: Structural equation models Structural equation models.
Factor Analysis ( 因素分析 ) Kaiping Grace Yao National Taiwan University
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
Statistical Analysis of Data1 of 38 1 of 42 Department of Cognitive Science Adv. Experimental Methods & Statistics PSYC 4310 / COGS 6310 MANOVA Multivariate.
Adjusted from slides attributed to Andrew Ainsworth
September 18-19, 2006 – Denver, Colorado Sponsored by the U.S. Department of Housing and Urban Development Conducting and interpreting multivariate analyses.
Lecture 12 Factor Analysis.
ANCOVA. What is Analysis of Covariance? When you think of Ancova, you should think of sequential regression, because really that’s all it is Covariate(s)
Applied Quantitative Analysis and Practices
Profile Analysis Intro and Assumptions Psy 524 Andrew Ainsworth.
Exploratory Factor Analysis. Principal components analysis seeks linear combinations that best capture the variation in the original variables. Factor.
Education 795 Class Notes Factor Analysis Note set 6.
Chapter 13.  Both Principle components analysis (PCA) and Exploratory factor analysis (EFA) are used to understand the underlying patterns in the data.
Advanced Statistics Factor Analysis, I. Introduction Factor analysis is a statistical technique about the relation between: (a)observed variables (X i.
Applied Quantitative Analysis and Practices LECTURE#19 By Dr. Osman Sadiq Paracha.
FACTOR ANALYSIS 1. What is Factor Analysis (FA)? Method of data reduction o take many variables and explain them with a few “factors” or “components”
Principal Component Analysis
MANOVA Lecture 12 Nuance stuff Psy 524 Andrew Ainsworth.
FACTOR ANALYSIS.  The basic objective of Factor Analysis is data reduction or structure detection.  The purpose of data reduction is to remove redundant.
Chapter 14 EXPLORATORY FACTOR ANALYSIS. Exploratory Factor Analysis  Statistical technique for dealing with multiple variables  Many variables are reduced.
FACTOR ANALYSIS & SPSS. First, let’s check the reliability of the scale Go to Analyze, Scale and Reliability analysis.
1 FACTOR ANALYSIS Kazimieras Pukėnas. 2 Factor analysis is used to uncover the latent (not observed directly) structure (dimensions) of a set of variables.
Chapter 12 REGRESSION DIAGNOSTICS AND CANONICAL CORRELATION.
Lecture 2 Survey Data Analysis Principal Component Analysis Factor Analysis Exemplified by SPSS Taylan Mavruk.
FACTOR ANALYSIS & SPSS.
Exploratory Factor Analysis
EXPLORATORY FACTOR ANALYSIS (EFA)
Analysis of Survey Results
Showcasing the use of Factor Analysis in data reduction: Research on learner support for In-service teachers Richard Ouma University of York SPSS Users.
Measuring latent variables
An introduction to exploratory factor analysis in IBM SPSS Statistics
Calculating Sample Size: Cohen’s Tables and G. Power
Advanced Data Preparation
© LOUIS COHEN, LAWRENCE MANION AND KEITH MORRISON
Measuring latent variables
Measuring latent variables
EPSY 5245 EPSY 5245 Michael C. Rodriguez
Principal Component Analysis
Factor Analysis.
Measuring latent variables
Presentation transcript:

Introduction to Multivariate Analysis of Variance, Factor Analysis, and Logistic Regression Rubab G. ARIM, MA University of British Columbia December 2006

Topics Multivariate Analysis of Variance (MANOVA) Multivariate Analysis of Variance (MANOVA) Factor Analysis Factor Analysis –Principal Component Analysis Logistic Regression Logistic Regression

MANOVA Extension of ANOVA Extension of ANOVA More than one dependent variable (DV) More than one dependent variable (DV) –Conceptual reason –Statistically related Compares the groups and tells whether there are group mean differences on the combination of the DVs Compares the groups and tells whether there are group mean differences on the combination of the DVs

Why not just conduct a series of ANOVAs? Risk of an inflated Type 1 error: Risk of an inflated Type 1 error: The more analyses you run, the more likely you are to find a significant result, even if in reality there are no differences between groups. If you choose to do so: Bonferroni adjustment--divide your alpha value.05 by the number of tests that you are intending to perform Bonferroni adjustment--divide your alpha value.05 by the number of tests that you are intending to perform

MANOVA: Pros and Cons MANOVA prevents the inflation of Type 1 error MANOVA prevents the inflation of Type 1 error Controls for correlation among a set of DVs by combining them Controls for correlation among a set of DVs by combining themHowever, A complex set of procedures A complex set of procedures Additional assumptions required Additional assumptions required

Example Research Question: Research Question: Do adolescent boys and girls differ in their problem behaviors? What you need? What you need? –One categorical IV (i.e., gender) –Two or more continuous DVs (e.g., depression, aggression, – etc.)

Example (cont’) What MANOVA does What MANOVA does –Tests the null hypothesis that the population means on a set of DVs do not vary across different levels of a grouping variable Assumptions Assumptions –sample size, normality, outliers, linearity, multicollinearity, homogeneity of variance-covariance matrices

Interpretation of the output Descriptive Statistics Descriptive Statistics –Check N values (more subjects in each cell than the number of DVs) Box’s Test Box’s Test –Checking the assumption of variance- covariance matrices Levene’s Test Levene’s Test –Checking the assumption of equality of variance

Interpretation (cont’) Multivariate tests Multivariate tests –Wilks’ Lambda (most commonly used) –Pillai’s Trace (most robust) (see Tabachnick & Fidell, 2007) Tests of between-subjects effects Tests of between-subjects effects –Use a Bonferroni Adjustment –Check Sig. column

Interpretation (cont’) Effect size Effect size –Partial Eta Squared: the proportion of the variance in the DV that can be explained by the IV (see Cohen, 1988) Comparing group means Comparing group means –Estimated marginal means Follow-up analyses Follow-up analyses (see Hair et al., 1998; Weinfurt, 1995)

Factor Analysis (FA) Not designed to test hypotheses Not designed to test hypotheses Data reduction technique Data reduction technique –Whether the data may be reduced to a smaller set of components or factors Used in the development and evaluation of tests and scales Used in the development and evaluation of tests and scales

Two main approaches in FA Exploratory factor analysis (EFA) Exploratory factor analysis (EFA) –Explore the interrelationships among a set of variables Confirmatory factor analysis (CFA) Confirmatory factor analysis (CFA) –Confirm specific hypotheses or theories concerning the structure underlying a set of variables

Principal Component Analysis (PCA) A technique similar to Factor Analysis in the sense that PCA also produces a smaller number of variables that accounts for most of the variability in the pattern or correlations A technique similar to Factor Analysis in the sense that PCA also produces a smaller number of variables that accounts for most of the variability in the pattern or correlationsHowever, Factor Analysis Factor Analysis –Mathematical model: only the shared variance in the variables is analyzed Principal Component Analysis Principal Component Analysis –All the variance in the variables are used

PCA or FA? If you are interested in a theoretical solution, use FA If you are interested in a theoretical solution, use FA If you want an empirical summary of your data set, use PCA If you want an empirical summary of your data set, use PCA (see Tabachnick & Fidell, 2001)

Steps involved in PCA Assessment of the suitability of the data Assessment of the suitability of the data –Sample size (see Stevens, 1996) –Strength of the relationship among the items an inspection of the correlation matrix r >.30 –Bartlett’s test of sphericity (p <.05) –Kaiser-Meyer Olkin (KMO) This index ranges from 0 to 1, with.6 suggested as the minimum value

Steps involved in PCA (cont’) Factor Extraction Factor Extraction –Determine the smallest number of factors that best represent the interrelations among the set of items –Various techniques (e.g., principal factor analysis, maximum likelihood factoring) –Determine the number of factors Kaiser’s criterion (eigenvalue > 1) Kaiser’s criterion (eigenvalue > 1) Scree test (plots each eigenvalue, find the point where the shape becomes horizontal) Scree test (plots each eigenvalue, find the point where the shape becomes horizontal)

Steps involved in PCA (cont’) Factor rotation and interpretation Factor rotation and interpretation –Orthogonal (uncorrelated) factor solutions Varimax is the most common technique –Oblique (correlated) factor solutions Direct Oblimin is the most common technique –Simple structure (Thurstone, 1947): each factor is represented by a number of strongly loading items

Example Research Question: Research Question: –What is the underlying factor structure of the Subjective Age Identity (SAI) scale? What you need What you need –A set of correlated continuous variables (i.e., items of the SAI scale) What PCA does What PCA does –Attempts to identify a small set of factors that represents the underlying relationships among a group of related variables (i.e., SAI items)

Example (cont’) Assumptions Assumptions –Sample size N > 150+ and a ratio of at least five cases for each of the items –Factorability of the correlation matrix r =.3 or greater; KMO ≥.6; Bartlett (p <.05) –Linearity –Outliers among cases

Interpretation of the output Is PCA appropriate? Is PCA appropriate? –Check Correlation Matrix –Check KMO and Bartlett’s test How many factors? Eigenvalue > 1 How many factors? Eigenvalue > 1 –Check the Total Variance Explained –Look at the Scree Plot

Interpretation (cont’) How many components are extracted? How many components are extracted? –Component Matrix –Rotated Component Matrix Look for the highest loading items on each of the component-this can be used to identify the nature of the underlying latent variable represented by each component

Logistic Regression Three types of regression Three types of regression –Bivariate –Multiple –Logistic* Relationships among variables Relationships among variables (NOT mean differences) One DV + 2 or more predictors or explanatory variables One DV + 2 or more predictors or explanatory variables *The DV is dichotomous *The DV is dichotomous *Core concept: Odds Ratio (OR) *Core concept: Odds Ratio (OR)

Logistic Regression Program A Program B Male Female50150 For males, the odds of watching Program A are: 200/100 (or 2 to 1). For females, the odds of watching Program A are: 50/150 (or 1 to 3). To obtain the ratio of the odds for gender relative to Program A: This OR = (2/1) / (1/3) = 6 >Males are six time more likely to be watching Program A.

Example Research Question: Research Question: Are adolescent girls more likely to have anxiety/depression? What you need? What you need? –One categorical IV (i.e., gender) –One dichotomous DV (non-depressed=0 and depressed = 1)

Interpretation of the output Nagelkerke R 2 Nagelkerke R 2 Is the model significant? Wald’s Test Wald’s Test At the parameter-level of inference, is the gender variable significant?

Selected References Pallant, J. (2004). SPSS survival manual: A step by step guide to data analysis using SPSS (2nd ed.). Maidenhead: Open University Press. Pallant, J. (2004). SPSS survival manual: A step by step guide to data analysis using SPSS (2nd ed.). Maidenhead: Open University Press. Pett, M. A., Lackey, N. R., Sullivan, J. J. (2003). Making sense of factor analysis: The use of factor analysis for instrument development in health care research. Thousand Oaks, CA: Sage. Tabachnick, B. G., & Fidell, L. S. (2001). Using multivariate statistics (4th.ed.). Boston: Allyn & Bacon. Tabachnick, B. G., & Fidell, L. S. (2001). Using multivariate statistics (4th.ed.). Boston: Allyn & Bacon.