Common Factor Analysis “World View” of PC vs. CF Choosing between PC and CF PAF -- most common kind of CF Communality & Communality Estimation Common Factor.

Slides:



Advertisements
Similar presentations
Automated Regression Modeling Descriptive vs. Predictive Regression Models Four common automated modeling procedures Forward Modeling Backward Modeling.
Advertisements

Factor Analysis Continued
Chapter Nineteen Factor Analysis.
Correlation & Regression Chapter 15. Correlation statistical technique that is used to measure and describe a relationship between two variables (X and.
Psychology 202b Advanced Psychological Statistics, II April 7, 2011.
MULTIPLE REGRESSION. OVERVIEW What Makes it Multiple? What Makes it Multiple? Additional Assumptions Additional Assumptions Methods of Entering Variables.
Principal Components An Introduction Exploratory factoring Meaning & application of “principal components” Basic steps in a PC analysis PC extraction process.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Principal Components An Introduction exploratory factoring meaning & application of “principal components” Basic steps in a PC analysis PC extraction process.
Chapter 4 Multiple Regression.
Power Analysis for Correlation & Multiple Regression Sample Size & multiple regression Subject-to-variable ratios Stability of correlation values Useful.
When Measurement Models and Factor Models Conflict: Maximizing Internal Consistency James M. Graham, Ph.D. Western Washington University ABSTRACT: The.
Determining the # Of PCs Remembering the process Some cautionary comments Statistical approaches Mathematical approaches “Nontrivial factors” approaches.
“Ghost Chasing”: Demystifying Latent Variables and SEM
The Terms that You Have to Know! Basis, Linear independent, Orthogonal Column space, Row space, Rank Linear combination Linear transformation Inner product.
Bivariate & Multivariate Regression correlation vs. prediction research prediction and relationship strength interpreting regression formulas process of.
Today Concepts underlying inferential statistics
LECTURE 16 STRUCTURAL EQUATION MODELING.
Chapter 7 Correlational Research Gay, Mills, and Airasian
Discriminant Analysis Testing latent variables as predictors of groups.
Review Guess the correlation. A.-2.0 B.-0.9 C.-0.1 D.0.1 E.0.9.
Multivariate Methods EPSY 5245 Michael C. Rodriguez.
Structural Equation Modeling Continued: Lecture 2 Psy 524 Ainsworth.
Factor Analysis Psy 524 Ainsworth.
Principal Components An Introduction
Objectives of Multiple Regression
Principal Components Analysis BMTRY 726 3/27/14. Uses Goal: Explain the variability of a set of variables using a “small” set of linear combinations of.
Near East University Department of English Language Teaching Advanced Research Techniques Correlational Studies Abdalmonam H. Elkorbow.
Discriminant Function Analysis Basics Psy524 Andrew Ainsworth.
MULTIPLE TRIANGLE MODELLING ( or MPTF ) APPLICATIONS MULTIPLE LINES OF BUSINESS- DIVERSIFICATION? MULTIPLE SEGMENTS –MEDICAL VERSUS INDEMNITY –SAME LINE,
Multiple Regression The Basics. Multiple Regression (MR) Predicting one DV from a set of predictors, the DV should be interval/ratio or at least assumed.
Psy 427 Cal State Northridge Andrew Ainsworth PhD.
Estimation Kline Chapter 7 (skip , appendices)
Advanced Correlational Analyses D/RS 1013 Factor Analysis.
6. Evaluation of measuring tools: validity Psychometrics. 2012/13. Group A (English)
Factor Analysis Psy 524 Ainsworth. Assumptions Assumes reliable correlations Highly affected by missing data, outlying cases and truncated data Data screening.
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
Correlation & Regression Chapter 15. Correlation It is a statistical technique that is used to measure and describe a relationship between two variables.
Adjusted from slides attributed to Andrew Ainsworth
Lecture 12 Factor Analysis.
Multivariate Analysis and Data Reduction. Multivariate Analysis Multivariate analysis tries to find patterns and relationships among multiple dependent.
Module III Multivariate Analysis Techniques- Framework, Factor Analysis, Cluster Analysis and Conjoint Analysis Research Report.
Applied Quantitative Analysis and Practices
Education 795 Class Notes Factor Analysis Note set 6.
Latent regression models. Where does the probability come from? Why isn’t the model deterministic. Each item tests something unique – We are interested.
Multivariate Data Analysis Chapter 3 – Factor Analysis.
Estimation Kline Chapter 7 (skip , appendices)
Advanced Statistics Factor Analysis, I. Introduction Factor analysis is a statistical technique about the relation between: (a)observed variables (X i.
FACTOR ANALYSIS 1. What is Factor Analysis (FA)? Method of data reduction o take many variables and explain them with a few “factors” or “components”
Principal Component Analysis
D/RS 1013 Discriminant Analysis. Discriminant Analysis Overview n multivariate extension of the one-way ANOVA n looks at differences between 2 or more.
FACTOR ANALYSIS.  The basic objective of Factor Analysis is data reduction or structure detection.  The purpose of data reduction is to remove redundant.
Chapter 14 EXPLORATORY FACTOR ANALYSIS. Exploratory Factor Analysis  Statistical technique for dealing with multiple variables  Many variables are reduced.
Chapter 12 REGRESSION DIAGNOSTICS AND CANONICAL CORRELATION.
Lecture 2 Survey Data Analysis Principal Component Analysis Factor Analysis Exemplified by SPSS Taylan Mavruk.
Principal Components & Common Factoring An Introduction
EXPLORATORY FACTOR ANALYSIS (EFA)
Week 3 Class Discussion.
Measuring latent variables
Reliability and Validity of Measurement
Measuring latent variables
I. Statistical Tests: Why do we use them? What do they involve?
Measuring latent variables
EPSY 5245 EPSY 5245 Michael C. Rodriguez
Multivariate Statistics
Factor Analysis BMTRY 726 7/19/2018.
Lecture 8: Factor analysis (FA)
Factor Analysis.
Measuring latent variables
Presentation transcript:

Common Factor Analysis “World View” of PC vs. CF Choosing between PC and CF PAF -- most common kind of CF Communality & Communality Estimation Common Factor Scores

World View of PC Analyses PC analysis is based on a very simple “world view” We measure variables The goal of factoring is data reduction determine the # of kinds of information in the variables build a PC for each R holds the relationships between the variables PCs are composite variables computed from linear combinations of the measured variables

World View of CF Analyses CF is based on a somewhat more complicated and “causal” world view Any domain (e.g., intelligence, personality) has some set of “latent constructs” A person’s “values” on these “latent constructs” causes their scores on any measured variable(s) any variable has two parts “common part” -- caused by values of the latent constructs” “unique part” -- not related to any latent construct (“error”)

World View of CF Analyses, cont the goal of factoring is to reveal the number and identify of these “latent constructs” R must be “adjusted” to represent the relationships between portions of the variables that are produced by the “latent constructs” represent the correlations between the “common” parts of the variables CFs are linear combinations of the “common” parts of the measured variables that capture the underlying constructs”

Example of CF world view “latent constructs” IQ Math Ability Reading Skill Social Skills “measures” adding, subtraction, multiplication vocabulary, reading speed, reading comprehension politeness, listening skills, sharing skills Each measure is “produced” by a weighted combination of the latent constructs, plus something unique to that measure... adding =.5*IQ +.8*Math + 0*Reading + 0*Social + U a subtraction =.5*IQ +.8*Math + 0*Reading + 0*Social + U s vocabulary =.5*IQ + 0*Math +.8*Reading + 0*Social + U v politeness=.4*IQ + 0*Math + 0*Reading +.8*Social + U p

Example of CF world view, cont When we factor these, we might find something like CF1 CF2 CF3 CF4 adding.4.6 subtraction.4.6 multiplication.4.6 vocabulary.4.6 reading speed.4.6 reading comp.4.6 politeness.3.6 listening skills.3.6 sharing skills.3.6 Name each “latent construct” that was revealed by this analysis

Principal Axis Analysis “Principal” again refers to the extraction process each successive factor is orthogonal and accounts for the maximum available covariance among the variables “Axis” tells us that the fact o rs are extracted from a “reduced” correlation matrix diagonals < 1.00 diagonals = the estimated “communality” of each variable reflecting that not all of the variance of that variable is “produced” by the set of “latent variables” So, factors extracted from the “reduced” R will reveal the latent variables

Which model to choose -- PC or PAF ? Traditionally... PC is used for “psychometric” purposes reduction of collinear predictor sets examination of the structure of “scoring systems” consideration of scales and sub-scales works with full R because composites will be computed from original variable scores not “common parts” CF is used for “theoretical” purposes identification of “underlying constructs” number and identity of “basic elements of behavior” The basis for “latent class” analyses of many kinds both measurement & structural models works with reduced R because it hold the “meaningful” part of the variables and their interrelationships The researcher selects the procedure based on their purpose for the factor analysis !!

Communality & Its Estimation The communality of a variable is the proportion of that variable’s variance that is produced by the common factors underlying the set of variables Common Estimations  (reliability coefficient) -- only the reliable part of the variable can be common largest r (or r 2 ) with another in the set -- at least that much is shared with other variables R 2 predicting that variable from all the others -- tells how much is shared with other variables Note how the definition shifts from “variance shared with the latent constructs” to “variance shared with the other variables in the set” !! 8)\

Communality & Its Estimation: How SPSS does it… Step 1: Perform a PC analysis extract # PCs from the full R matrix Step 2: Perform 1 st PAF Iteration Use R 2 predicting each variable from others -- put in diagonal of R extract same # PAFs from that reduced R matrix compute (output) variable communalities Step 3: Perform 2 nd PAF Iteration use variable (output) communalities from last PAF step as estimated (input) communalities -- put in diagonals of R extract same # PAFs from that reduced R matrix compute (output) variable communalities Compare estimated (input) and computed (output) variable communalities Additional Steps: Iterate to convergence of estimated (input) & computed (output) variable communalities

Communality & Its Estimation How SPSS does it…, cont Huh?!!? The idea is pretty simple (and elegant) … If the communality estimates are correct, then they will be returned from the factor analysis ! So, start with a “best guess” of the communalities, and iterate until the estimates are stable Note: This takes advantages of the “self-correcting” nature of this iterative process the initial estimates have very little effect on the final communalities (R 2 really easy to calculate) starting with the PC communalities tends to work quickly Note: This process assumes the latent constructs are adequately represented by the variable set !!

Problems estimating communalities in a CF analysis “failure to converge” usually this can be solved by increasing the number of iterations allowed (=1000) “Heywood case”  λ > 1.00 During iteration communality estimates can become larger than 1.00 However no more than “all” of a variable’s variance can be common variance! Usual solutions… Use the solution from the previous iteration Drop the offending variable If other variables are “threatening to Heywood” consider aggregating them together into a single variable

Common Factor Scores The “problem” is that common factors can only be computed as combinations of the “common parts” of the variables Unfortunately, we can’t separate each person’s score on each variable into the “common” and “unique” part So, common factor scores have to be “estimated” Good news -- the procedure used by SPSS works well and is well accepted since CF is done for “theory testing” or to “reveal latent constructs” rather than for “psychometric purposes” scores for CFs are not used as often as are PC scores

Maximum Likelihood method of Common Factoring Both PAF & ML are “common factor” extractions they both seek to separate the “common” vs. “unique” portion of each variable’s variance and include only the common in R they both require communality estimates they both iterate communality input estimates & output computations until these two converge, though the process for computing estimates is somewhat different which is taken as evidence that the communality estimates are accurate and so, S extracted using those estimates describes the factor structure of R PAF factors are extracted to derive S that will give the best reproduction of variance in sampled R matrix ML factors are extracted to derive S that is most likely to represent population S % reproduce the population R

Maximum Likelihood method of Common Factoring If assumptions of interval measurement and normal distribution are well-met, ML works somewhat better than PAF & vice versa ML is an extraction technique – the rotational techniques discussed for PC and PAF all apply to ML factors ML is a common factoring technique – issue of factor score “estimation ” are the same as for PAF Proponents of ML exploratory factoring emphasize … ML estimation procedures are most the common in confirmatory factoring, latent class measurement, structural models & the generalized linear model ML estimation permits an internally consistent set of significance tests – e.g., # factors decisions.