Presentation is loading. Please wait.

Presentation is loading. Please wait.

Structural Equation Modeling

Similar presentations


Presentation on theme: "Structural Equation Modeling"— Presentation transcript:

1 Structural Equation Modeling
Verbal d1 Math Analytic d2 d3 d4 d9 d10 d11 d12 d6 d5 d7 d8 Structural Equation Modeling

2 What is SEM? Combines measurement models of CFA with goals of multiple regression analysis to allow the prediction of latent variables from other latent variables. Simultaneous regression equations Modeling latent variables from observed variables Estimate parameters of the measurement model & structural model Comparison between implied covariance matrix & observed covariance matrix

3 Advantages of SEM Testing multiple relationships at a time
Multiple independent and dependent variables can be accommodated (DVs can even related to one another) Examining latent variables (but must link them to manifest variables) Specifying measurement error in the model Allows enhanced model fit No assumption of uncorrelated errors (although by default errors uncorrelated  need to change to allow correlated errors)

4 Model Testing in SEM Measurement model Path model Structural model
Also known as confirmatory factor analysis Tests the relationship between the indicators and the latent variables they are supposed to measure Path model Tests the relationships between the exogenous and endogenous variables without the measurement model specified Structural model Tests the relationships between the exogenous and endogenous variables with the measurement model specified

5 Measurement Model 12  1 X1 X2 X3 X4 1 2 3 4 41 31 21 11  2
5 6 7 8 82 72 62 52 12

6 Path Model X1 X4 X6 X2 X5 X3

7 Structural Model 1 2 3 4 1 2 3 4 X1 X2 X3 X4 y1 y2 y3 y4 11
21 31 41 11 21 31 41 11  1 1 21 21  2 52 82 62 72 y5 y6 y7 y8  5  6  7  8

8 SEM Lingo Exogenous variable – Construct that acts only as a predictor or cause in a model, an IV. Not predicted by anything in the model (A) Endogenous variable – Construct that is an outcome variable in at least one causal relationship, a DV, also mediators (B, C, D) C B A D

9 SEM Lingo: Constructs and Indicators
Exogenous constructs/latent variables are called Ksis – represented by ξ Endogenous constructs/latent variables are called etas –represented by η Exogenous indicator/manifest variable - X Endogenous indicator/manifest variable - Y

10 SEM Lingo: Constructs and Indicators
Y X Y ξ η X η Y X Y Y Ksis Etas

11 SEM Lingo Nonrecursive – relationship is reciprocal in path diagram
Recursive – relationships not reciprocal Nested Models – Models that have same constructs but differ in the number and type of causal relationships represented (i.e., parameters estimated) In nested models, one model is a subset of the other

12 Measurement Model Matrices
Lambda-X (ΛX) Loadings of exogenous indicators; tells how you get from the manifest Xs to the latent Xs Lambda-Y (ΛY) Loadings of endogenous indicators; tells how you get from manifest Ys to latent Ys Theta-delta (θd) Errors of the exogenous indicators, the manifest X variables Theta-epsilon (θε) Errors of the endogenous indicators, the manifest Y variables 8 matrices are used in SEM. Four of them are for the measurement model and four for the structural model

13 Example   Intelligence School Performance GMAT IQ Test GPA # Pubs LX
LY LY LX TE TE TD TD d d

14 Structural Model Matrices
Beta (B) Relationships of endogenous constructs to endogenous constructs; how DVs cause each other Gamma (Г) Relationships of exogenous constructs to endogenous constructs; how IVs cause DVs *Phi (Φ) Correlations among latent exogenous constructs; correlations among the IVs Psi (Ψ) Residuals from prediction of latent endogenous constructs; Tells whether residuals of prediction are correlated * Phi matrix is relevant in CFA because is allows us to determine whether we are looking at an ablique or orthogonal rotation (are the latent variables correlated?) By default, the values of phi will be free (correlation allowed – oblique). Some of the newer output will provide a correlation matrix of all etas and ksis rather than the phi matrix (this will have the phi matrix submsumed under it).

15 Example   d Gamma Gamma Beta Phi Gamma GPA # Pubs GMAT IQ Test d
Work Performance Sup Rating Peer Motivation Sup Report Self d Gamma Gamma Beta Phi Gamma School Performance GPA # Pubs Intelligence GMAT IQ Test d Gamma Psi – Resid

16 Assumptions in SEM Observations are independent
Respondents are randomly sampled In maximum likelihood estimation, multivariate normality assumption Continuous variables (using correlations) except when use: Polychoric correlation matrix Tetrachoric correlation matrix Polyserial correlation matrix Biserial correlation matrix Polychoric correlation matrix – two ordinal variables, 3+ categories Tetrachoric correlation matrix – two binary variables Polyserial correlation matrix – 1 metric, 1 ordinal measure with 3+ categories Biserial correlation matrix – 1 binary measure, 1 metric measure

17 SEM Sample Size Requirements
Absolute minimum = number of covariances or correlations in the matrix Typical min = 5 respondents/parameter estimated, 10/parameter preferred When not MV normal – 15 respondents per parameter ML estimation Can use as few as 50, but recommended Ideal n = 200 Estimation techniques other than ML tend to require larger sample sizes

18 Estimation Procedures
In SEM, goal of estimation is to minimize error between the observed and reproduced values in VCV matrix  choose parameter estimates that increase likelihood of reproducing VCV matrix Regr: min Σ(y – y’)2 SEM: min (obs VCV – repr VCV) Note this is analygous to LSE in regression

19 Estimation Procedures
Unweighted Least Squares (OLS) Used in regression, but not in SEM OLS is scale invariant only if the errors of measurement are uncorrelated  this is an assumption in regression but not in SEM Assumes MV normality

20 SEM Estimation Procedures
Generalized Least Squares (GLS) Used in SEM You take the least squares and weight it with a VCV matrix  yields a scale-free estimation procedure Assumes MV Normality

21 SEM Estimation Procedures
Maximum Likelihood Estimation (ML) Weights least squares estimates with VCV matrix; updates VCV matrix each iteration Assumes ML normality As increase sample, GLS = ML Finds parameter estimates that maximize the probability of the data Most commonly used and default estimation procedure in LISREL

22 SEM Estimation Procedures
Weighted Least Squares Makes no assumptions about distribution Need huge sample (n = 500+) No assumption of ML normality required In practice, WLS not really used. ML and GLS are robust against assumptions of multivariate normality

23 Steps in Conducting SEM
Draw a picture of your model including both your latent and manifest variables Test the fit of your measurement model. Adjust as needed to enhance fit of measurement model Once measurement model fits, test fit of structural model Modify structural model involves doing exploratory SEM  not recommended

24 Anderson & Gerbing Two Step Approach
Step 1. Adequacy of Measurement Model Test measurement model; allow all latent variables to correlate Adjust meas model as needed to enhance fit If fit of measurement model is poor, don’t test structural model Fit of structural model is a necessary but not sufficient condition for the fit of structural model Step 2. Adequacy of Structural Model

25 Hair - Stages in SEM Develop a theoretically based model
Construct a path diagram of causal relationships Convert path diagram into set of structural and measurement models Choose the input matrix type and estimate the proposed model Assess the identification of the model Evaluate goodness-of-fit criteria Interpret and modify the model

26 Develop Model & Path Diagram
product factors price-based relationship usage satisfaction with company

27 Convert path diagram to structural and measurement models
X3 X4 Y1 product factors price-based relationship usage satisfaction w/ company Y3 X1 X2 Y4 Y2 X5 X6

28 Choose Input Matrix & Estimate Model
Careful with missing data! No “Missing data” correlation matrix Choice of correlation matrix or VCV matrix Correlation matrix yields standardized weights from +1 to -1 VCV matrix is better for validating causal relationships

29 Analyzing Correlation versus Covariance Matrices
SEM models are based on the decomposition of covariance matrices, not correlation matrices. The solutions hold, strictly speaking, for the analysis of covariance matrices. To the extent that the solution depends on the scale of the variables, analyses based on covariance matrices and correlation matrices can differ.

30 Model Identification Degrees of freedom (df) are related to the number of parameter estimates Model df must be > or = 0 Just-identified model/saturated model: df = 0; perfect model fit *Over-identified model: df > 0 because more information in the matrix than the number of parameters estimates Under-identified model: df < 0 because model has more parameters estimated than information available. Can’t run due to infinite solutions Compare number of data points to number of parameters where the number of data points is equal to the # of variances and covariances. Overidentification when more data points than parameters to be estimated – this is required to run SEM If too many parameters are estimated, to reduce the number of parameters estimated we need to: Fixing – setting a parameter to a specific value Constraining – setting a parameter equal to another parameter Deleting To establish a scale in a factor: Fix the variance of the factor to 1 Fix the regression coefficient from the factor to one of the measured variables to 1

31 Identification Problems
Often results from a large number of parameters estimated compared to number of correlations provided  too few degrees of freedom Solution: Estimate fewer parameters

32 Model Fit Fit of model denoted by two things:
Small residuals Nonsignificant difference between original VCV matrix and reconstructed VCV matrix To assess model it, SEM provides numerous goodness of fit indices Different indices assess fit in different ways

33 Types of Goodness of Fit Indices
Absolute Fit Measures Overall model fit, no adjustment for overfitting Incremental Fit Measures Compare proposed model fit to another model specified by researcher Parsimonious Fit Measures “Adjust” model to provide comparison between models with differing numbers of estimated coefficients

34 Specific Goodness of Fit Indices
Absolute Fit Measures Chi2 (2) Goodness-of-fit (GFI) Root Mean Square Error of Approximation (RMSEA) Root Mean Square Residual (RMR) Incremental Fit Measures Adjusted-goodness-of-fit (AGFI) Normed Fit Index (NFI) Parsimonious Fit Measures Parsimony Normed Fit Index (PNFI) Parsimony Goodness of Fit Index (PGFI)

35 Chi2 - 2 A “badness of fit” measure
Represents the extent to which the observed and reproduced correlation matrices differ High power (high n size) increases 2 so that it is significant penalized for large n 2 only look at when n.s. 2 difference test – compares nested models Most practical use of chi2

36 SEM - Degrees of Freedom
df = number of known pieces of information – unknowns to be estimated Important in model fit if estimate fewer paths, fit will reduce just by chance (almost all paths not 0 just due to chance) Best possible model fit saturated model in which all the links are estimated by default Number of data points versus number of parameters to be estimated. The # of data points in SEM is the # of variances and covariances in

37 Goodness of fit index (GFI)
The quality of the original model and its ability to reproduce the actual variance-covariance matrix is more easily gauged by GFI This index is similar to R2 in multiple regression This index tells us how much better our model compared to the null model Want higher values, > or = .90

38 SEM - Degrees of Freedom
Manifest A1 Manifest A1 Latent A Latent A Manifest A2 Manifest A2 Manifest B1 Manifest B1 Latent B Latent B Manifest B2 Manifest B2 The second model will fit better (smaller chi2) merely because an additional parameter is estimated. The question of interest is: Is the fit significantly better than it would be without this additional link? Manifest C1 Manifest C1 Latent C Latent C Manifest C2 Manifest C2 Larger 2 - Worse Fit Smaller 2 - Better Fit

39 Root Mean Square Error of Approximation (RMSEA)
A normed index with rules of thumb Prefer a RMSEA < or = .05

40 Root Mean Squared Residual (RMR)
Want this value to be small, < or = .05 is ideal, < or = .1 is probably good Not a normed fit index; size of residuals influenced by variance of variables involved This is just the square root of the residuals. Smaller residuals indicate better fit

41 Adjusted Goodness of fit index (AGFI)
Adjusted goodness-of-fit index (AGFI) was created to account for increases in fit due to chance Similar to the adjusted R2 in multiple regression Can be negative Less sensitive to changes in df than is PNFI/PGFI

42 Normed Fit Index (NFI) NFI compares fit of the null model to fit of the theoretical model Large values of NFI are best (ideally > or = .9) Criticism of NFI: Comparing the model fit to a model of nothing. Is this meaningful?

43 Parsimony Normed Fit Index (PNFI)
Problem: Most fit indices increase just by estimating more parameters (free more links to be estimated) Just identified model  perfect fit PNFI penalizes you for lack of parsimony No clear benchmark for what is “good” – best to compare between models

44 Parsimony Goodness of Fit Index (PGFI)
Gets smaller as you increase the number of paths in the model No clear benchmark for what is “good” – best to compare between models

45 Some Common Rules of Thumb for Model Fit
Test or Index Good Fit Acceptable Fit Chi-Square Goodness of Fit p > .20 p > .05 GFI .95 .90 AGFI .80 RMR Depends on scale. Closer to 0 is better RMSEA < .05 < .08

46 Model Testing Strategies with SEM
Model Confirmation Single model tested to fit or not fit Problem with “confirmation bias” – many possible models fit *Competing Models Strategy Compares competing models for best fit Nested models should be used Exploratory SEM/Model development Capitalizes on Chance

47 Testing Competing Nested Models
Comparing the fit of hypothesized and alternative models that have the same constructs but differ in number of parameters estimated Models must be nested within one another to compare them  use 2 difference test to compare Typical nested models involve deleting or adding a single path

48 Testing Nested Models MODEL 1 A B C MODEL 2 A B C These two models are nested within one another because they differ only in the addition of a single link in the second model

49 Likelihood Ratio Test Problem: sensitive to sample size Solution: CFI
changes in CFI less than -.01 Cheung & Rensvold (2002) Structural Equation Modeling, 9, CFI is the comparative fit index. It is a modified version of the NFI (an incremental fit index) that is less sensitive to sample size – should be interpreted same as NFI but better for model comparison because sample size won’t affect it

50 Model Modification: Exploratory SEM
Common management practice not to do this  don’t edit structural models  a confirmatory technique (but edit meas model OK) t values Tells us where deleting a path would enhance model fit (paths with n.s. t values)

51 Model Modification: Exploratory SEM
Modification Indices Suggest where paths could be added to increase fit Represents the degree to which 2 would decrease if you added a path To justify change due to MI Fairly large Justifiable Impact model fit Not central to theory When using modification indices to enhance fit, you should only consider MIs that are > 10. Start with the largest MI and add that link – then consider the effects on the model as a whole and see if any more additions would help. A high MI should be associated with a high residual and the residuals should decrease once you free the parameter to be estimated there. Using MIs to modify model is exploratory SEM and involves capitalization on chance, thus some folks think you should not do this. If you choose to, add all the links to the model and then delete links you don’t want based on t test values (< 2) that reflect n.s. parameter estimates.

52 In LISREL, the modification indices are the changes in the goodness-of-fit c2 that would result from setting that parameter free. Modification Indices

53 In LISREL, the modification indices are the changes in the goodness-of-fit c2 that would result from setting that parameter free.

54 SEM with LISREL Traditional LISREL language SIMPLIS language* PRELIS
Uses matrix language Specify whether aspects of matrix are free (FR) or fixed (FI) SIMPLIS language* More recent LISREL language More user-friendly syntax PRELIS Prepares raw data for use in LISREL - generates correlation matrix or VCV matrix to input

55 Running LISREL Program
Title line Input Specification Model Specification Path Diagram Output Specification

56 Exercise: CFA Model Draw the diagram  1  2 X1 X2 X3 X4 X5 X6 X7 X8

57 Exercise: CFA Model Draw the diagram 12  1 X1 X2 X3 X4 1 2 3 4
41 31 21 11  2 X5 X6 X7 X8 5 6 7 8 82 72 62 52 12

58 Input Specification: SIMPLIS
Title line: Exercise Observed variables: X1 X2 X3 X4 X5 X6 X7 X8 Covariance matrix: 1.65 Sample size: 200 Latent variables: LAT1 LAT2 You could input a correlation matrix rather than a VCV matrix or have it get raw data.

59 LISREL/SIMPLIS Code (continued)
RELATIONSHIPS X1 = 1*LAT1 X2 = LAT1 X3 = LAT1 X4 = LAT1 X5 = 1*LAT2 X6 = LAT2 X7 = LAT2 X8 = LAT2 LAT1 = LAT2 LISREL OUTPUT: SS SC EF AD = OFF PRINT RESIDUALS PATH DIAGRAM END OF PROBLEM AD = Admissiaaibility check. Stops after a default number of iterations (20) if can’t converge on a solution. Usually we shut this off because LISREL can almost never converge with it turned on. Once shut off can specify another number of iterations you would like to allow if you’d like.

60 LISREL/SIMPLIS Code: Estimation Techniques
Default estimation technique is ML Other Techniques available: Generalized least squares (GLS) Unweighted least squares (ULS) Weighted least squares (WLS) Other techniques with this syntax Method of estimation: GLS

61 LISREL/SIMPLIS Code: LISREL Output
SS Print standardized solution SC Print completely standardized solution EF Print total & indirect effects, their standard errors & t values VA Print variances and covariances FS Print factor scores regression PC Print correlations of parameter estimates PT Print technical information

62 Example: Confirmatory Factor Analysis

63 Verbal d1 Math Analytic d2 d3 d4 d9 d10 d11 d12 d6 d5 d7 d8

64 Tests of significance for parameter estimates: t values.

65 Parameter estimates

66 CHI-SQUARE WITH 51 DEGREES OF FREEDOM = 55.50 (P = 0.31)
ESTIMATED NON-CENTRALITY PARAMETER (NCP) = 4.50 90 PERCENT CONFIDENCE INTERVAL FOR NCP = (0.0 ; 26.77) MINIMUM FIT FUNCTION VALUE = 0.11 POPULATION DISCREPANCY FUNCTION VALUE (F0) = 90 PERCENT CONFIDENCE INTERVAL FOR F0 = (0.0 ; 0.054) ROOT MEAN SQUARE ERROR OF APPROXIMATION (RMSEA) = 0.013 90 PERCENT CONFIDENCE INTERVAL FOR RMSEA = (0.0 ; 0.032) P-VALUE FOR TEST OF CLOSE FIT (RMSEA < 0.05) = 1.00 EXPECTED CROSS-VALIDATION INDEX (ECVI) = 0.22 90 PERCENT CONFIDENCE INTERVAL FOR ECVI = (0.21 ; 0.26) ECVI FOR SATURATED MODEL = 0.31 ECVI FOR INDEPENDENCE MODEL = 3.98 CHI-SQUARE FOR INDEPENDENCE MODEL WITH 66 DEGREES OF FREEDOM = INDEPENDENCE AIC = MODEL AIC = SATURATED AIC = INDEPENDENCE CAIC = MODEL CAIC = SATURATED CAIC = ROOT MEAN SQUARE RESIDUAL (RMR) = 0.028 STANDARDIZED RMR = 0.028 GOODNESS OF FIT INDEX (GFI) = 0.98 ADJUSTED GOODNESS OF FIT INDEX (AGFI) = 0.97 PARSIMONY GOODNESS OF FIT INDEX (PGFI) = 0.64 NORMED FIT INDEX (NFI) = 0.97 NON-NORMED FIT INDEX (NNFI) = 1.00 PARSIMONY NORMED FIT INDEX (PNFI) = 0.75 COMPARATIVE FIT INDEX (CFI) = 1.00 INCREMENTAL FIT INDEX (IFI) = 1.00 RELATIVE FIT INDEX (RFI) = 0.96 CRITICAL N (CN) =

67 Hypothesized Model Fit Statistics
CHI-SQUARE WITH 51 DEGREES OF FREEDOM = (P = 0.31) (This test models the variances and covariances as implied by the parameter expectations) CHI-SQUARE FOR INDEPENDENCE MODEL WITH 66 DEGREES OF FREEDOM = (This test only models the variances of the variables and assumes all covariances are 0) GOODNESS OF FIT INDEX (GFI) = 0.98 ADJUSTED GOODNESS OF FIT INDEX (AGFI) = 0.97

68 CORRELATION MATRIX TO BE ANALYZED
V V V V M M2 V V V V M M M M R R R R M M R R R R4 M M R R R R

69 FITTED COVARIANCE MATRIX
V V V V M M2 V V V V M M M M R R R R M M R R R R4 M M R R R R Fitted covariance matrix takes into account the scale of measurement

70 FITTED RESIDUALS V1 V2 V3 V4 M1 M2
V V V V M M M M R R R R M M R R R R4 M M R R R R Residuals should be small if fit is good. Fitted residuals depend on the unit of measurement

71 STANDARDIZED RESIDUALS
V V V V M M2 V V V V M M M M R R R R M M R R R R4 M M R R R R Standardized residuals is a residual divided by the estimated standard error – adjusted for the scale of measurement

72 The chosen model fits the data quite well. How would other models do
The chosen model fits the data quite well. How would other models do? A complete confirmatory analysis would not only test the preferred model but also examine alternative models to assess how easily they could account for the data. To the extent that reasonable alternatives exist, the preferred model must be considered with more caution.

73 Alternative Measurement Model 1
Verbal Math Analytic d1 d2 d3 d4 d5 d6 d7 d8 d9 d10 d11 d12

74 Parameter Estimates

75

76 Alternative Model 1 Fit Indices
CHI-SQUARE WITH 54 DEGREES OF FREEDOM = (P = 0.0) CHI-SQUARE FOR INDEPENDENCE MODEL WITH 66 DEGREES OF FREEDOM = GOODNESS OF FIT INDEX (GFI) = 0.93 ADJUSTED GOODNESS OF FIT INDEX (AGFI) = 0.90

77 Alternative Measurement Model 2
F1 d1 d2 d3 d4 d5 d6 d7 d8 d9 d10 d11 d12

78 Parameter Estimates

79

80 Alternative Model 2 Fit Indices
CHI-SQUARE WITH 54 DEGREES OF FREEDOM = (P = 0.0) CHI-SQUARE FOR INDEPENDENCE MODEL WITH 66 DEGREES OF FREEDOM = GOODNESS OF FIT INDEX (GFI) = 0.73 ADJUSTED GOODNESS OF FIT INDEX (AGFI) = 0.61

81 Summary Fit Statistics
Model Chi2 df GFI AGFI Hyp Alt Alt Note: Our hypothesized measurement model is the best fit!!

82 Example Structural Model Testing

83 Hypothesized Model e e e e Home Aspire e e e e Ability Achieve e e e
Family Income e Ed. Aspirations e Father Education Home Aspire e Occ. Aspirations e Mother Education e e Verbal Ability Verbal Achieve Ability Achieve e Quant. Ability e Quant. Achieve e

84 Syntax to Specify Hypothesized Structural Model
RELATIONSHIPS faminc = 1*home faed = home moed = home verbab = 1*ability quantab = ability edasp = 1*aspire ocasp = aspire verach = 1*achieve quantach = achieve home = ability aspire = home ability achieve = home ability Syntax to Specify Hypothesized Structural Model Just the second part of the syntax which does not reflect the beginning stuff to enter data Note the 1* denotes that the factor loading has been fixed to one – this sets the latent variable to the scale of measurement used by the manifest variable.

85 COVARIANCE MATRIX TO BE ANALYZED
edasp ocasp verach quantach faminc faed edasp ocasp verach quantach faminc faed moed verbab quantab moed verbab quantab moed verbab quantab

86 GOODNESS OF FIT STATISTICS
CHI-SQUARE WITH 21 DEGREES OF FREEDOM = 57.17 (P = ) ROOT MEAN SQUARE ERROR OF APPROXIMATION (RMSEA) = 0.093 CHI-SQUARE FOR INDEPENDENCE MODEL WITH 36 DEGREES OF FREEDOM = ROOT MEAN SQUARE RESIDUAL (RMR) = 0.047 STANDARDIZED RMR = 0.048 GOODNESS OF FIT INDEX (GFI) = 0.94 ADJUSTED GOODNESS OF FIT INDEX (AGFI) = 0.87

87 Parameter estimates

88 T tests

89 Parameter estimates

90

91

92

93

94

95

96

97 STANDARDIZED SOLUTION
LAMBDA-Y aspire achieve edasp ocasp verach quantach LAMBDA-X home ability faminc faed moed verbab quantab Lambda Y - Loadings of endogenous indicators; tells how you get from the manifest Ys to the latent Ys Lambda X - Loadings of exogenous indicators; tells how you get from the manifest Xs to the latent Xs

98 STANDARDIZED SOLUTION
BETA aspire achieve aspire achieve GAMMA home ability aspire achieve CORRELATION MATRIX OF ETA AND KSI aspire achieve home ability aspire achieve home ability PSI Beta (B) Relationships of endogenous constructs to endogenous constructs; how DVs cause each other Gamma (Г) Relationships of exogenous constructs to endogenous constructs; how IVs cause DVs *Phi (Φ) Correlations among latent exogenous constructs (ksis); correlations among the IVs Psi (Ψ) Residuals from prediction of latent endogenous constructs; Tells whether residuals of prediction are correlated Replace phi – Correlations of ksis and etas Ksi - Exogenous constructs/latent variables Eta - Endogenous constructs/latent variables

99 FITTED COVARIANCE MATRIX
edasp ocasp verach quantach faminc faed edasp ocasp verach quantach faminc faed moed verbab quantab moed verbab quantab moed verbab quantab Covariance matrix on the indicator variables

100 FITTED RESIDUALS edasp ocasp verach quantach faminc faed
edasp ocasp verach quantach faminc faed moed verbab quantab moed verbab quantab moed verbab quantab SUMMARY STATISTICS FOR FITTED RESIDUALS SMALLEST FITTED RESIDUAL = MEDIAN FITTED RESIDUAL = LARGEST FITTED RESIDUAL =

101 STANDARDIZED RESIDUALS
edasp ocasp verach quantach faminc faed edasp ocasp verach quantach faminc faed moed verbab quantab moed verbab quantab moed verbab quantab SUMMARY STATISTICS FOR STANDARDIZED RESIDUALS SMALLEST STANDARDIZED RESIDUAL = MEDIAN STANDARDIZED RESIDUAL = LARGEST STANDARDIZED RESIDUAL = Note highest residual between faed and moed – if you looked at MI would also see MI

102 STANDARDIZED RESIDUALS
QPLOT OF STANDARDIZED RESIDUALS x x x N x O x R x x . M x x A xx L x x x x Q x x U x A xx N xx T x x I * L * E x S x x x . x STANDARDIZED RESIDUALS

103 Alternative Model 1 Home Achieve Ability Aspire e Family Income
Father Education Mother Education Verbal Ability Quant. Ability Ed. Aspirations Occ. Aspirations Verbal Achieve Quant. Achieve e

104 Syntax to Specify Alternative Structural Model
RELATIONSHIPS faminc = 1*home faed = home moed = home verbab = 1*ability quantab = ability edasp = 1*aspire ocasp = aspire verach = 1*achieve quantach = achieve home = ability aspire = home ability achieve = home ability Let the errors for faed and moed correlate Syntax to Specify Alternative Structural Model

105 CHI-SQUARE WITH 20 DEGREES OF FREEDOM = 19.17 (P = 0.51)
ROOT MEAN SQUARE ERROR OF APPROXIMATION (RMSEA) = 0.0 90 PERCENT CONFIDENCE INTERVAL FOR RMSEA = (0.0 ; 0.058) CHI-SQUARE FOR INDEPENDENCE MODEL WITH 36 DEGREES OF FREEDOM = ROOT MEAN SQUARE RESIDUAL (RMR) = 0.015 STANDARDIZED RMR = 0.015 GOODNESS OF FIT INDEX (GFI) = 0.98 ADJUSTED GOODNESS OF FIT INDEX (AGFI) = 0.95

106

107

108

109

110

111

112

113

114

115

116 STANDARDIZED SOLUTION
LAMBDA-Y aspire achieve edasp ocasp verach quantach LAMBDA-X home ability faminc faed moed verbab quantab

117 BETA aspire achieve aspire achieve GAMMA home ability aspire achieve CORRELATION MATRIX OF ETA AND KSI aspire achieve home ability aspire achieve home ability PSI

118 COVARIANCE MATRIX TO BE ANALYZED
edasp ocasp verach quantach faminc faed edasp ocasp verach quantach faminc faed moed verbab quantab moed verbab quantab moed verbab quantab

119 FITTED COVARIANCE MATRIX
edasp ocasp verach quantach faminc faed edasp ocasp verach quantach faminc faed moed verbab quantab moed verbab quantab moed verbab quantab

120 FITTED RESIDUALS edasp ocasp verach quantach faminc faed
edasp ocasp verach quantach faminc faed moed verbab quantab moed verbab quantab moed verbab quantab

121 STANDARDIZED RESIDUALS
edasp ocasp verach quantach faminc faed edasp ocasp verach quantach faminc faed moed verbab quantab moed verbab quantab moed verbab quantab SUMMARY STATISTICS FOR STANDARDIZED RESIDUALS SMALLEST STANDARDIZED RESIDUAL = MEDIAN STANDARDIZED RESIDUAL = LARGEST STANDARDIZED RESIDUAL =

122 STANDARDIZED RESIDUALS
QPLOT OF STANDARDIZED RESIDUALS x x x N x O x R xx M x A xx L * * Q * U x. x A xx N * T x x I x L * E x S x x x x STANDARDIZED RESIDUALS

123

124 Original Model Modified Model BETA aspire achieve BETA
aspire achieve GAMMA home ability aspire achieve BETA aspire achieve aspire achieve GAMMA home ability aspire achieve Original Model Modified Model

125 Chi2 difference (1) = 38.00 Original Model Modified Model
Chi (df = 21) (df = 20) RMSEA RMR SRMR GFI AGFI Chi2 difference (1) = 38.00


Download ppt "Structural Equation Modeling"

Similar presentations


Ads by Google