Confirmatory Factor Analysis Psych 818 DeShon. Purpose ● Takes factor analysis a few steps further. ● Impose theoretically interesting constraints on.

Slides:



Advertisements
Similar presentations
1 Regression as Moment Structure. 2 Regression Equation Y =  X + v Observable Variables Y z = X Moment matrix  YY  YX  =  YX  XX Moment structure.
Advertisements

Structural Equation Modeling. What is SEM Swiss Army Knife of Statistics Can replicate virtually any model from “canned” stats packages (some limitations.
SEM PURPOSE Model phenomena from observed or theoretical stances
Structural Equation Modeling Using Mplus Chongming Yang Research Support Center FHSS College.
Structural Equation Modeling
The Multiple Regression Model.
Hypothesis Testing Steps in Hypothesis Testing:
Hypothesis: It is an assumption of population parameter ( mean, proportion, variance) There are two types of hypothesis : 1) Simple hypothesis :A statistical.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
SOC 681 James G. Anderson, PhD
Structural Equation Modeling
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
The Simple Linear Regression Model: Specification and Estimation
Factor Analysis Ulf H. Olsson Professor of Statistics.
1 Chapter 3 Multiple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
Multivariate Data Analysis Chapter 11 - Structural Equation Modeling.
When Measurement Models and Factor Models Conflict: Maximizing Internal Consistency James M. Graham, Ph.D. Western Washington University ABSTRACT: The.
“Ghost Chasing”: Demystifying Latent Variables and SEM
Structural Equation Modeling
Measurement Models and CFA Ulf H. Olsson Professor of Statistics.
GRA 6020 Multivariate Statistics Confirmatory Factor Analysis Ulf H. Olsson Professor of Statistics.
Aaker, Kumar, Day Seventh Edition Instructor’s Presentation Slides
G Lecture 51 Estimation details Testing Fit Fit indices Arguing for models.
Linear and generalised linear models
LECTURE 16 STRUCTURAL EQUATION MODELING.
The General (LISREL) SEM model Ulf H. Olsson Professor of statistics.
Stages in Structural Equation Modeling
Multiple Sample Models James G. Anderson, Ph.D. Purdue University.
Confirmatory factor analysis
AM Recitation 2/10/11.
Aaker, Kumar, Day Ninth Edition Instructor’s Presentation Slides
Structural Equation Modeling 3 Psy 524 Andrew Ainsworth.
Moderation in Structural Equation Modeling: Specification, Estimation, and Interpretation Using Quadratic Structural Equations Jeffrey R. Edwards University.
Structural Equation Modeling
CJT 765: Structural Equation Modeling Class 7: fitting a model, fit indices, comparingmodels, statistical power.
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
ROB CRIBBIE QUANTITATIVE METHODS PROGRAM – DEPARTMENT OF PSYCHOLOGY COORDINATOR - STATISTICAL CONSULTING SERVICE COURSE MATERIALS AVAILABLE AT:
CJT 765: Structural Equation Modeling Class 8: Confirmatory Factory Analysis.
1 General Structural Equations (LISREL) Week 1 #4.
Confirmatory Factor Analysis Psych 818 DeShon. Construct Validity: MTMM ● Assessed via convergent and divergent evidence ● Convergent – Measures of the.
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
Assessing Hypothesis Testing Fit Indices
Measures of Fit David A. Kenny January 25, Background Introduction to Measures of Fit.
Advanced Statistics Factor Analysis, II. Last lecture 1. What causes what, ξ → Xs, Xs→ ξ ? 2. Do we explore the relation of Xs to ξs, or do we test (try.
Multivariate Statistics Confirmatory Factor Analysis I W. M. van der Veld University of Amsterdam.
Measurement Models: Identification and Estimation James G. Anderson, Ph.D. Purdue University.
G Lecture 81 Comparing Measurement Models across Groups Reducing Bias with Hybrid Models Setting the Scale of Latent Variables Thinking about Hybrid.
SEM Basics 2 Byrne Chapter 2 Kline pg 7-15, 50-51, ,
Assessing Hypothesis Testing Fit Indices Kline Chapter 8 (Stop at 210) Byrne page
Chap 6 Further Inference in the Multiple Regression Model
CJT 765: Structural Equation Modeling Class 8: Confirmatory Factory Analysis.
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
SEM Model Fit: Introduction David A. Kenny January 12, 2014.
Examples. Path Model 1 Simple mediation model. Much of the influence of Family Background (SES) is indirect.
ALISON BOWLING CONFIRMATORY FACTOR ANALYSIS. REVIEW OF EFA Exploratory Factor Analysis (EFA) Explores the data All measured variables are related to every.
Confirmatory Factor Analysis of Longitudinal Data David A. Kenny December
The general structural equation model with latent variates Hans Baumgartner Penn State University.
Evaluation of structural equation models Hans Baumgartner Penn State University.
Chapter 14 EXPLORATORY FACTOR ANALYSIS. Exploratory Factor Analysis  Statistical technique for dealing with multiple variables  Many variables are reduced.
Chapter 17 STRUCTURAL EQUATION MODELING. Structural Equation Modeling (SEM)  Relatively new statistical technique used to test theoretical or causal.
The SweSAT Vocabulary (word): understanding of words and concepts. Data Sufficiency (ds): numerical reasoning ability. Reading Comprehension (read): Swedish.
Testing the Equality of Means and Variances in Over-Time Data David A. Kenny.
Measures of Fit David A. Kenny.
Advanced Statistical Methods: Continuous Variables
Structural Equation Modeling using MPlus
Chapter 15 Confirmatory Factor Analysis
CJT 765: Structural Equation Modeling
Writing about Structural Equation Models
Confirmatory Factor Analysis
Testing Causal Hypotheses
Presentation transcript:

Confirmatory Factor Analysis Psych 818 DeShon

Purpose ● Takes factor analysis a few steps further. ● Impose theoretically interesting constraints on the model and examine the resulting fit of the model with the observed data ● Used to evaluate theoretical measurement structures ● Provides tests and indices to evaluate fit

Purpose ● CFA model is constructed in advance – specifies the number of (latent) factors – specifies the pattern of loadings on the factors – that specifies the pattern of unique variances specific to each observation – measurement errors may be correlated - Yuck! – factor loadings can be constrained to be zero (or any other value) – covariances among latent factors can be estimated or constrained – multiple group analysis is possible ● Can TEST if these constraints are consistent with the data

CFA Model & Notation Where: x = (q  1) vector of indicator/manifest variable  = (q  n) matrix of factor loadings – lambda  = (n  1) vector of latent constructs (factors) – ksi or xi  = (q  1) vector of errors of measurement – delta

CFA Example ● Measures for positive emotions,  1 : – x 1 = Happiness, x 2 =Pride ● Measures for negative emotions  2 : – x 3 = Sadness, x 4 =Fear ● Model

CFA Example HappyPrideSadFear Pos Neg δ1δ1 δ2δ2 δ3δ3 δ4δ4 1.0 More on the 1.0s later

CFA Model Matrices

More Matrices Theta-delta Phi

Model Fitting ● The specified model results in an implied variance-covariance matrix, Σ θ

Parameter Estimation ● Maximum Likelihood Estimation – Assumes multivariate normality – Iterative procedure – Requires starting values ● F ML = Tr(SC-1) - n + ln(det(C)) – ln(det(S)) – S is the sample variance-covariance matrix – C is the implied variance-covariance matrix

Model Identification ● Definition: – The set of parameters  ={ , ,  } is not identified if there exists  1  2 such that  (  1 )=  (  2 ).

1 Factor, 2 indicators ● Not Identified! ● Population covariance matrix ● Implied Covariance Matrix ● Solutions

1 Factor, 3 Indicators ● Just-Identified; always fits perfectly

1 Factor, 3 Indicators ● Just-Identified; always fits perfectly

Identification Rules ● Number of free parameters: t  ½ q (q+1) ● Three-Indicator Rule – Exactly 1 non zero element per row of  – 3 or more indicators per factor –  Diagonal– uncorrelated errors ● Two-Indicator Rule –  ij  0 for at least one pair i, j, i  j – Exactly 1 non-zero element per row of  – 2 or more indicators per factor –  Diagonal – uncorrelated errors

Scaling the latent variables ● The scale/metric of the latent variable is not determinant – Factor loadings and variances can take on any value unless the metric is specified ● Must impose a model constraint to yield a meaningful scale ● Two Constraints are possible: – Fix a loading to 1.0 from each factor to one of its indicators ● Latent scale takes on the metric of the constrained indicator – Fix the latent variance to 1.0 ● Yields a standardized latent variable

Scaling the Latent Variables δ1δ1 δ2δ2 δ3δ3 δ4δ4 δ5δ5 δ6δ6 x1x1 x2x2 x3x3 x4x4 x5x5 x 16 ξ2ξ2 ξ1ξ1 1 1 δ1δ1 δ2δ2 δ3δ3 δ4δ4 δ5δ5 δ6δ6 x1x1 x2x2 x3x3 x4x4 x5x5 ξ2ξ2 ξ1ξ1 1 1 Fix Path Fix Variances

Scaling the Latent Variables

Covariance or Correlation matrix ● Cudeck shows that the covariance matrix is prefered for CFA ● Use of a correlation matrix can result in many problems – Incorrect parameter estimates – Incorrect standard errors – Incorrect test statistics

Parameter Evaluation – Local Fit ● Parameter Estimate/SE = Z-statistic ● Standard interpretation: – if |Z| > 2, then “significant” ● Consider both statistical and scientific value of including a variable in the model ● Significance testing in CFA: – Not usually interesting in determining if loadings are equal to zero – Might be interested in testing whether or not covariance between factors is zero.

Goodness of Fit – Global Fit ● Absolute indices – derived from the fit of the obtained and implied covariance matrices and the ML minimization function. – Chi-square & functions of it ● Relative fit indices – Relative to baseline “worst-fitting” model ● Adjusted fit indices – Relative to number of parameters in model – Parsimony

Chi Square:  2 ● F ML *(n-1) ● Sensitive to sample size – The larger the sample size, the more likely the rejection of the model and the more likely a Type II error (rejecting something true). In very large samples, even tiny differences between the observed model and the perfect-fit model may be found significant. – Informative when sample sizes are relatively small ( ) ● Chi-square fit index is also very sensitive to violations of the assumption of multivariate normality

Relative Fit Indices ● Fit indices that provide information relative to a baseline model – a model in which all of the correlations or covariances are zero – Very poor fit – Termed Null or Independence model ● Permits evaluation of the adequacy of the target model

Goodness of Fit Index (GFI) ● % of observed covariances explained by the covariances implied by the model – Ranges from 0-1 ● Biased – Biased up by large samples – Biased downward when degrees of freedom are large relative to sample size ● GFI is often higher than other fit indices – Not commonly used any longer

Normed Fit Index (Bentler & Bonnet, 1980) ● Compare to a Null or Independence model – a model in which all of the correlations or covariances are zero ● % of total covariance among observed variables explained by target model when using the null (independence) model as baseline - Hu & Bentler,1995 ● Not penalized for a lack of parsimony ● Not commonly used anymore

Non-Normed Fit Index - Tucker & Lewis, 1973 ● Penalizes for fitting too many parameters ● May be greater than 1.0 – If so, set to 1.0

Comparative Fit Index - Bentler, 1989; 1990 ● Based on noncentrality parameter for chi- square distribution ● Indicates reduction in model misfit of a target model relative to a baseline (independence) model ● Can be greater than 1.0 or less than 0.0 – If so, set to 1.0 or 0.0

Root Mean Square Error of Approximation (RMSEA) ● Discrepancy per degree of freedom ● RMSEA  0.05  Close fit ● 0.05 <RMSEA  0.08  Reasonable fit ● RMSEA > 0.1  Poor fit

Standardized Root Mean Square Residual (SRMR) ● Standardized difference between the observed covariance and predicted covariance ● A value of zero indicates perfect fit ● This measure tends to be smaller as sample size increases and as the number of parameters in the model increases.

General fit standards ● NFI – = acceptable; above.95 is good – NFI positively correlated with sample size ● NNFI – = acceptable; above.95 is good – NFI and NNFI not recommended for small sample sizes ● CFI – = acceptable; above.95 is good – No systematic bias with small sample size

General fit standards ● RMSE – Should be close to zero – 0.0 to 0.05 is good fit – 0.05 to 0.08 is moderate fit – Greater than.10 is poor fit ● SRMR – Less than.08 is good fit ● Hu & Bentler – SRMR .08 AND (CFI .95 OR RMSEA .06)

Nested Model Comparisons ● Test between equivalent models except for a subset of parameters that are freely estimated in one and fixed in the other ● Difference in –2LL is distributed as chi- square variate – Each model has a  2 value based upon a certain degree of freedom – If models are nested (ie., identical but M 2 deletes one parameter found in M 1 ), significance of ‘increment’ or ‘decrement’ in fit =   2 2 with df = df 1 – df 2

Modification indexes ● If the model does not fit well, modification indices may be used to guide respecification (ie. how to improve the model) ● For CFA, the only sensible solutions are to add direct paths from construct to indicator to drop paths ● Rarely is it reasonable to let residuals covary as many times suggested by the output ● Respecify the model or drop factors or indicators

Reliability in CFA ● Standard reliability assumes tau-equivalence. ● Can estimate reliability in CFA with no restrictions – Congeneric measures

Equivalent models ● Equivalent models exist for almost all models ● Most quant people say you should evaluate alternative models ● Most people don't do it ● For now, be aware that there are alternative models that will fit as well as your model

Heirarchical CFA Quality of life for adolescents: Assessing measurement properties using structural equation modelling Lynn B. Meuleners, Andy H. Lee, Colin W. Binns & Anthony Lower Quality of Life Research 12: 283–290, 2003.

Heirarchical CFA Depression (CES-D ) Positive Affect Negative Affect Somatic Symptoms HappyEnjoyBotheredBluesDepressedSadMindEffortSleep.795 a.882 a.810 a Model Fit Statistics: N= 868, χ 2 (26)= , p<.001, SRMR=.055, IFI=.976 a Second-order loadings were set equal for empirical identification. All loadings significant at p <.001.

Example...AMOS