Chapter 13.  Both Principle components analysis (PCA) and Exploratory factor analysis (EFA) are used to understand the underlying patterns in the data.

Slides:



Advertisements
Similar presentations
Structural Equation Modeling. What is SEM Swiss Army Knife of Statistics Can replicate virtually any model from “canned” stats packages (some limitations.
Advertisements

Structural Equation Modeling
Factor Analysis Continued
Exploratory Factor Analysis
© LOUIS COHEN, LAWRENCE MANION & KEITH MORRISON
Lecture 7: Principal component analysis (PCA)
Psychology 202b Advanced Psychological Statistics, II April 7, 2011.
WENDIANN SETHI SPRING 2011 SPSS ADVANCED ANALYSIS.
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 14 Using Multivariate Design and Analysis.
Factor Analysis Ulf H. Olsson Professor of Statistics.
Principal Components An Introduction Exploratory factoring Meaning & application of “principal components” Basic steps in a PC analysis PC extraction process.
Factor Analysis Research Methods and Statistics. Learning Outcomes At the end of this lecture and with additional reading you will be able to Describe.
1 Carrying out EFA - stages Ensure that data are suitable Decide on the model - PAF or PCA Decide how many factors are required to represent you data When.
Goals of Factor Analysis (1) (1)to reduce the number of variables and (2) to detect structure in the relationships between variables, that is to classify.
Education 795 Class Notes Factor Analysis II Note set 7.
Multivariate Methods EPSY 5245 Michael C. Rodriguez.
Factor Analysis Psy 524 Ainsworth.
Social Science Research Design and Statistics, 2/e Alfred P. Rovai, Jason D. Baker, and Michael K. Ponton Factor Analysis PowerPoint Prepared by Alfred.
What is Factor Analysis?
2 nd Order CFA Byrne Chapter 5. 2 nd Order Models The idea of a 2 nd order model (sometimes called a bi-factor model) is: – You have some latent variables.
SEM: Basics Byrne Chapter 1 Tabachnick SEM
Psy 427 Cal State Northridge Andrew Ainsworth PhD.
MGMT 6971 PSYCHOMETRICS © 2014, Michael Kalsher
Advanced Correlational Analyses D/RS 1013 Factor Analysis.
Applied Quantitative Analysis and Practices
Factor Analysis Psy 524 Ainsworth. Assumptions Assumes reliable correlations Highly affected by missing data, outlying cases and truncated data Data screening.
Thursday AM  Presentation of yesterday’s results  Factor analysis  A conceptual introduction to: Structural equation models Structural equation models.
Factor Analysis Revealing the correlational structure among variables Understanding & Reducing Complexity.
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
Multigroup Models Byrne Chapter 7 Brown Chapter 7.
Introduction to Multivariate Analysis of Variance, Factor Analysis, and Logistic Regression Rubab G. ARIM, MA University of British Columbia December 2006.
CFA: Basics Beaujean Chapter 3. Other readings Kline 9 – a good reference, but lumps this entire section into one chapter.
Explanatory Factor Analysis: Alpha and Omega Dominique Zephyr Applied Statistics Lab University of Kenctucky.
Lecture 12 Factor Analysis.
SEM Basics 2 Byrne Chapter 2 Kline pg 7-15, 50-51, ,
Multivariate Analysis and Data Reduction. Multivariate Analysis Multivariate analysis tries to find patterns and relationships among multiple dependent.
Applied Quantitative Analysis and Practices
Exploratory Factor Analysis. Principal components analysis seeks linear combinations that best capture the variation in the original variables. Factor.
Education 795 Class Notes Factor Analysis Note set 6.
Exploratory Factor Analysis Principal Component Analysis Chapter 17.
Principle Component Analysis and its use in MA clustering Lecture 12.
Department of Cognitive Science Michael Kalsher Adv. Experimental Methods & Statistics PSYC 4310 / COGS 6310 Factor Analysis 1 PSYC 4310 Advanced Experimental.
Multivariate Data Analysis Chapter 3 – Factor Analysis.
Factor Analysis I Principle Components Analysis. “Data Reduction” Purpose of factor analysis is to determine a minimum number of “factors” or components.
Advanced Statistics Factor Analysis, I. Introduction Factor analysis is a statistical technique about the relation between: (a)observed variables (X i.
Applied Quantitative Analysis and Practices LECTURE#19 By Dr. Osman Sadiq Paracha.
FACTOR ANALYSIS 1. What is Factor Analysis (FA)? Method of data reduction o take many variables and explain them with a few “factors” or “components”
CFA Model Revision Byrne Chapter 4 Brown Chapter 5.
 Seeks to determine group membership from predictor variables ◦ Given group membership, how many people can we correctly classify?
Principal Component Analysis
FACTOR ANALYSIS.  The basic objective of Factor Analysis is data reduction or structure detection.  The purpose of data reduction is to remove redundant.
Multivariate statistical methods. Multivariate methods multivariate dataset – group of n objects, m variables (as a rule n>m, if possible). confirmation.
Chapter 14 EXPLORATORY FACTOR ANALYSIS. Exploratory Factor Analysis  Statistical technique for dealing with multiple variables  Many variables are reduced.
FACTOR ANALYSIS & SPSS. First, let’s check the reliability of the scale Go to Analyze, Scale and Reliability analysis.
Data Screening. What is it? Data screening is very important to make sure you’ve met all your assumptions, outliers, and error problems. Each type of.
CFA: Basics Byrne Chapter 3 Brown Chapter 3 (40-53)
Lecture 2 Survey Data Analysis Principal Component Analysis Factor Analysis Exemplified by SPSS Taylan Mavruk.
FACTOR ANALYSIS & SPSS.
Exploratory Factor Analysis
EXPLORATORY FACTOR ANALYSIS (EFA)
Question So, I’ve done my factor analysis.
Analysis of Survey Results
Factor analysis Advanced Quantitative Research Methods
Measuring latent variables
Measuring latent variables
Descriptive Statistics vs. Factor Analysis
Measuring latent variables
EPSY 5245 EPSY 5245 Michael C. Rodriguez
Principal Component Analysis
Measuring latent variables
Presentation transcript:

Chapter 13

 Both Principle components analysis (PCA) and Exploratory factor analysis (EFA) are used to understand the underlying patterns in the data

 They group the variables into “factors” or “components” that are the processes that created the high correlations between variables.

 Exploratory factor analysis (EFA) – describe the data and summarize it’s factors  First step with research/data set  Confirmatory factor analysis (CFA) – already know latent factors – therefore, used to confirm relationship between factors and variables used to measure those factors.  Structural equation modeling

 Mathwise – summarizes patterns of correlations and reduce the correlations of variables into components/factors  Data reduction

 A popular use for both PCA and EFA is for scale development.  You can determine which questions best measure what you are trying to assess.  That way you can shorten your scale from 100 questions to maybe 15.

 Regression on crack  Creates linear combinations (regression equations) of the variables > which then is transposed into a component/factor

 Interpretation – as with clustering/scaling, one main problem with PCA/EFA is the interpretation.  A good analysis is explainable / make sense

 How do you know that this solution is the best solution?  There isn’t quite a good way to know if it’s a good solution like regression  Loads of rotation options

 EFA is usually a hot mess  As with every other type of statistical analysis we discuss, EFA has a certain type of research design associated with it.  Not a last resort on messy data.  AND often researchers do not apply the best established rules and therefore end up with results you don’t know what they mean.

 Observed correlation matrix – the correlations between all of the variables  Akin to doing a bivariate correlation chart  Reproduced correlation matrix – correlation matrix created from the factors.

 Residual correlation matrix – the difference between the original and reduced correlation matrix  You want this to be small for a good fitting model

 Factor rotation – process by which the solution is made “better” (smaller residuals) without changing the mathematical properties.

 Factor rotation – orthogonal – holds all the factors as uncorrelated (!!) Factor 1 Factor 2 Factor 1 Factor 2

 Factor rotation – orthogonal – varimax is the most common  Loading matrix – correlations between the variables and factors  Interpret the loading matrix  But – how many times in life are things uncorrelated?

 Factor rotation – oblique – factors are allowed to be correlated when they are rotated Factor 1 Factor 2 Factor 1 Factor 2

 Factor correlation matrix – correlations among the factors  Structure matrix – correlations between factors and variables  Pattern matrix – unique correlation between each factor and variables (no overlap which is allowed with rotation)  Similar to pr  Interpret pattern matrix

 Factor rotation – oblique rotations – oblimin, promax  You’ll know what type of rotation you’ve chosen by the output you get…

 EFA = produces factors  Only the shared variance and unique variance is analyzed  PCA = produces components  All the variance in the variables is analyzed

 EFA – factors are thought to cause variables, the underlying construct is what creates the scores on each variable  PCA – components are combinations of correlated variables, the variables cause the components

 How many variables?  You want several variables or items because if you only include 5, you are limited in the correlations that are possible AND the number of factors  Usually there’s about 10 (that could be expensive if you have to pay for your measures…)

 Sample size  The number one complaint about PCA and EFA is the sample size.  It is a make/break point in publications  Arguments abound what’s best.

 Sample size  100 is the lowest scrape by amount  200 is generally accepted as ok  300+ is the safest bet

 Missing data  PCA/EFA does not do missing data  Estimate the score, or delete it.

 Normality – multivariate normality is assumed  Its ok if they aren’t quite normal, but makes it easier to rotate when they are

 Linearity – correlations are linear! We expect there to linearity.

 Outliers - since this is regression and correlation – then outliers are still bad.  Zscores and mahalanobis

 PCA – multicollinearity = no big deal.  EFA – multicollinearity = delete or combine one of the overlapping variables.

 Unrelated variables (outlier variables) – only load on one factor – need to be deleted for a rerun of EFA.

 Dataset contains a bunch of personality characteristics  PCA – how many components do we expect?  EFA – how many factors do we expect?

 For PCA make sure this screen says “Principle components”  One leading problem with EFA is that people use Principle components math! Eek!  Ask for a scree plot  Pick a number of factors/let it pick**

 Communalities – how much variance of the variable is accounted for by the components.

 Eigenvalue box – remember eigenvalues are a mathematical way to rearrange the variance into clusters.  This box tells you how much variance each one of those “clusters”/eigenvalues account for.

 Scree plot – plots the eigenvalues

 Component matrix – the loading of each variable on each component.  You want them to load highly on components  BUT only on one component or it’s all confusing.  What’s high? .300 is a general rule of thumb

 Choose max likelihood or unweighted least squares

 Varimax – orthogonal rotation  Oblimin – oblique rotation

 Oblique vs Orthogonal?  Why why why use orthogonal?  Don’t force things to be uncorrelated when they don’t have to be!  If it’s truly uncorrelated oblique will give you the exact same results as orthogonal.

 How many factors?  Scree plot/eigenvalues  Look for the big drop  How much does a bootstrap analysis suggest (aka parallel analysis)?  Don’t just do how many eigenvalues over one (kaiser) all by itself

 Same boxes – then structure and pattern matrix  Interpret pattern matrix.  Loadings higher than.300

 Free little program that you can do factor analysis with…  Lots more rotation options  Other types of correlation options  Gives you more goodness of fit tests  Since SPSS doesn’t give you any!

 First read the data  You can save the data as space delimited from SPSS  You have to know the number of lines and columns

 Configure – select options you want  Types of correlations  Pearson for normally distributed continuous data sets  Polychloric for dichotomous data sets

 Parallel analysis or parallel bootstraps makes rotation easiest and quickest  Also crashes less  Number of factors  ULS/ML = EFA  PCA = PCA

 Rotations – you got a LOT of options. Good luck.  Compute!

 GOODNESS OF FIT STATISTICS  Chi-Square with 64 degrees of freedom = (P = )  Chi-Square for independence model with 91 degrees of freedom =  Non-Normed Fit Index (NNFI; Tucker & Lewis) = 0.94  Comparative Fit Index (CFI) = 0.96  Goodness of Fit Index (GFI) = 0.99  Adjusted Goodness of Fit Index (AGFI) = 0.98  Want these to be high!  Root Mean Square of Residuals (RMSR) =  Expected mean value of RMSR for an acceptable model = (Kelly's criterion)  Want these to be low!

 Preacher and MacCallum (2003)  Repairing Tom Swift’s Factor Analysis Machine  If you want to do EFA the right way, quote these people.