Structural Equation Modeling

Slides:



Advertisements
Similar presentations
1 Regression as Moment Structure. 2 Regression Equation Y =  X + v Observable Variables Y z = X Moment matrix  YY  YX  =  YX  XX Moment structure.
Advertisements

Structural Equation Modeling. What is SEM Swiss Army Knife of Statistics Can replicate virtually any model from “canned” stats packages (some limitations.
SEM PURPOSE Model phenomena from observed or theoretical stances
Structural Equation Modeling Using Mplus Chongming Yang Research Support Center FHSS College.
General Structural Equation (LISREL) Models
Structural Equation Modeling
Confirmatory Factor Analysis
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Ch11 Curve Fitting Dr. Deshi Ye
Linear Regression and Binary Variables The independent variable does not necessarily need to be continuous. If the independent variable is binary (e.g.,
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 14 Using Multivariate Design and Analysis.
Chapter 10 Simple Regression.
Chapter 4 Multiple Regression.
Multivariate Data Analysis Chapter 11 - Structural Equation Modeling.
Multivariate Data Analysis Chapter 4 – Multiple Regression.
CHAPTER 4 ECONOMETRICS x x x x x Multiple Regression = more than one explanatory variable Independent variables are X 2 and X 3. Y i = B 1 + B 2 X 2i +
1 1 Slide 統計學 Spring 2004 授課教師:統計系余清祥 日期: 2004 年 5 月 4 日 第十二週:複迴歸.
“Ghost Chasing”: Demystifying Latent Variables and SEM
Structural Equation Modeling
Topics: Regression Simple Linear Regression: one dependent variable and one independent variable Multiple Regression: one dependent variable and two or.
Chapter 11 Multiple Regression.
LECTURE 16 STRUCTURAL EQUATION MODELING.
Factor Analysis Ulf H. Olsson Professor of Statistics.
Correlation and Regression Analysis
Structural Equation Modeling Intro to SEM Psy 524 Ainsworth.
Simple Linear Regression Analysis
Discriminant Analysis Testing latent variables as predictors of groups.
Relationships Among Variables
Lecture 5 Correlation and Regression
Stages in Structural Equation Modeling
Structural Equation Modeling Continued: Lecture 2 Psy 524 Ainsworth.
Introduction to Linear Regression and Correlation Analysis
Chapter 11 Simple Regression
Structural Equation Modeling 3 Psy 524 Andrew Ainsworth.
SEM Analysis SPSS/AMOS
Discriminant Function Analysis Basics Psy524 Andrew Ainsworth.
Structural Equation Modeling (SEM) With Latent Variables James G. Anderson, Ph.D. Purdue University.
1 1 Slide © 2007 Thomson South-Western. All Rights Reserved Chapter 13 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
CJT 765: Structural Equation Modeling Class 7: fitting a model, fit indices, comparingmodels, statistical power.
1 1 Slide © 2008 Thomson South-Western. All Rights Reserved Chapter 15 Multiple Regression n Multiple Regression Model n Least Squares Method n Multiple.
Introduction to Linear Regression
SEM: Basics Byrne Chapter 1 Tabachnick SEM
Random Regressors and Moment Based Estimation Prepared by Vera Tabakova, East Carolina University.
CJT 765: Structural Equation Modeling Highlights for Quiz 2.
CJT 765: Structural Equation Modeling Class 8: Confirmatory Factory Analysis.
Regression Regression relationship = trend + scatter
Measurement Models: Exploratory and Confirmatory Factor Analysis James G. Anderson, Ph.D. Purdue University.
Chapter 13 Multiple Regression
Controlling for Baseline
Multivariate Statistics Confirmatory Factor Analysis I W. M. van der Veld University of Amsterdam.
Measurement Models: Identification and Estimation James G. Anderson, Ph.D. Purdue University.
SEM: Basics Byrne Chapter 1 Tabachnick SEM
SEM Basics 2 Byrne Chapter 2 Kline pg 7-15, 50-51, ,
CJT 765: Structural Equation Modeling Class 8: Confirmatory Factory Analysis.
ALISON BOWLING CONFIRMATORY FACTOR ANALYSIS. REVIEW OF EFA Exploratory Factor Analysis (EFA) Explores the data All measured variables are related to every.
1 1 Slide The Simple Linear Regression Model n Simple Linear Regression Model y =  0 +  1 x +  n Simple Linear Regression Equation E( y ) =  0 + 
Evaluation of structural equation models Hans Baumgartner Penn State University.
SEM with AMOS1 Christian A. Klöckner. 2 Agenda Expectations What is Structural Equation Modelling? What is AMOS? A simple regression A multiple regression.
Chapter 17 STRUCTURAL EQUATION MODELING. Structural Equation Modeling (SEM)  Relatively new statistical technique used to test theoretical or causal.
The SweSAT Vocabulary (word): understanding of words and concepts. Data Sufficiency (ds): numerical reasoning ability. Reading Comprehension (read): Swedish.
Advanced Statistical Methods: Continuous Variables
Structural Equation Modeling using MPlus
CJT 765: Structural Equation Modeling
Correlation, Regression & Nested Models
CJT 765: Structural Equation Modeling
Structural Equation Modeling
Simple Linear Regression
Causal Relationships with measurement error in the data
Structural Equation Modeling
Presentation transcript:

Structural Equation Modeling Karl L. Wuensch Dept of Psychology East Carolina University

Nomenclature AKA Two sets of variables “causal modeling.” “analysis of covariance structure.” Two sets of variables Indicators – measured (observed, manifest) variables – diagramed within rectangles Latent variables – factors – diagramed within ellipses

Causal Language Indicators and latent variables may be classified as “independent” or “dependent” Even if no variables are manipulated Based on the causal model being tested. In the diagram, indicators and latent variables may be connected by arrows One-headed = unidirectional causal flow Two-headed = direction not specified

Paths “Dependent” variables have one-headed arrows pointing to them. “Independent” variables do not. “Dependent” variables also have residuals (are not perfectly predicted by the independents) Called errors (e) for observed variables Disturbances (d) for latent variables

Two Models Measurement Model – how the measured variables are related to the latent variables. Structural Model – how the latent variables are related to each other.

Two Variance Covariance Matrices The sample matrix – computed from the sample data. The estimated population matrix – estimated from the model. 2 test of null that the model fits the data well. More useful are goodness of fit estimators

Sample Size Need at least 200 cases even for a simple model. Rule of thumb: at least 10 cases per estimated parameter.

Assumptions & Problems Multivariate normality. Linear relationships But can include polynomial components A singular matrix may crash the program Multicollinearity can be a problem

Simple Example from T&F The “Independent” Variables are shaded.

Regression Parameters Regression coefficients for the paths May be “fixed” to value 0 (no path) or 1 Or estimated from the data ().

Variance/Covariance Parameters Variances/Covariances of the “independent” variables May be estimated or fixed, to 1 or to the variance of a “marker” measured variable (set to 1 the path to the marker). Variances for “dependent” latent variables are usually fixed to the variance of one of the measured variables (set to 1 the path to that measured variable).

Model Identification A model is “identified” if there is a unique solution for each of the estimated parameters. Determine the number of input data points (values in the sample variance/covariance matrix). This is Where m = number of measured variables.

Model Identification For T&F’s simple model, 5(6)/2 = 15 data points.

Model Identification If the number of data points = the number of parameters to be estimated, the model is “just identified,” or “saturated,” and the fit will be perfect. If there are fewer data points than parameters to be estimated, the model is “under identified” and the analysis is kaput.

The “Over Identified” Model The number of input data points exceeds the number of parameters to be estimated. This is the desired situation. For T&F’s simple model, count the number of asterisks in the diagram. I count 11. 15 input data points, 11 parameters to estimate  we have an over identified model.

Eleven Parameters (*) The “Independent” Variables are shaded.

Identification of the Measurement Model Should Be OK if Only one latent variable, at least three indicators, errors not correlated. Two or more latent variables, each has at least three indicators, errors not correlated, each indicator loads on only one latent variable, latent variables are allowed to covary.

Identification of the Measurement Model Should Be OK if Two or more latent variables, one has only two indicators, errors are not correlated, each indicator loads on only one latent variable, none of the latent variables has a variance or covariance of zero.

Identification of the Structural Model May Be OK if None of the latent DVs predicts another latent DV, or if one does, it is recursive (unidirectional) and the disturbances are not correlated If there are nonrecursive relationships (an arrow from A to B and from B to A), hire an expert in SEM.

Error in Identification If there is a problem, the software will throw an error. The software may suggest a way to reach identification. You must tinker with the model to make it identified.

Estimation Maximum Likelihood most common An iterative procedure used to maximize fit between the sample var/cov matrix and the estimated population var/cov matrix. Generalized Least Squares estimation has also fared well in Monte Carlo comparisons of techniques.

Modifying and Comparing Models May simplify a model by deleting one or more parameters in it. The simplified model is nested with the previous model. Calculate difference 2 = difference between the two model’s Chi-Squares. df = number of parameters deleted.

LM and Wald Tests Lagrange Multiplier Test Wald Test Would fit be improved by estimating a parameter that is currently fixed? Wald Test Would fixing this parameter significantly reduce the fit? Available in SAS Calis, not in Amos

Reliability of Measured Variables Assume that the variance in the measured variable is due to variance in the latent variable (the “true scores”) + random error. Reliability = true variance divided by (true and error variance). Thus, estimated reliability = the r2 between measured variable and latent variable.