Presentation is loading. Please wait.

Presentation is loading. Please wait.

Mixture Modeling Chongming Yang Research Support Center FHSS College.

Similar presentations


Presentation on theme: "Mixture Modeling Chongming Yang Research Support Center FHSS College."— Presentation transcript:

1 Mixture Modeling Chongming Yang Research Support Center FHSS College

2 Mixture of Distributions

3

4 Classification Techniques Latent Class Analysis (categorical indicators) Latent Profile Analysis (continuous Indicators) Finite Mixture Modeling (multivariate normal variables) …

5 Integrate Classification Models into Other Models Mixture Factor Analysis Mixture Regressions Mixture Structural Equation Modeling Growth Mixture Modeling Multilevel Mixture Modeling

6 Disadvantages of Multi-steps Practice Multistep practice – Run classification model – Save membership Variable – Model membership variable and other variables Disadvantages – Biases in parameter estimates – Biases in standard errors Significance Confidence Intervals

7 Latent Class Analysis (LCA) Setting – Latent trait assumed to be categorical – Trait measured with multiple categorical indicators – Example: drug addiction, Schizophrenia Aim – Identify heterogeneous classes/groups – Estimate class probabilities – Identify good indicators of classes – Relate covariates to Classes

8 Graphic LCA Model Categorical Indicators u: u 1, u 2,u 3, …u r Categorical Latent Variable C: C =1, 2, …, or K

9 Probabilistic Model Assumption: Conditional independence of u so that interdependence is explained by C like factor analysis model An item probability Joint Probability of all indicators

10 LCA Parameters Number of Classes -1 Item Probabilities -1

11 Class Means (Logit)

12 Latent Class Analysis with Covariates

13 Posterior Probability (membership/classification of cases)

14 Estimation Maximum Likelihood estimation via Expectation-Maximization algorithm – E (expectation) step: compute average posterior probabilities for each class and item – M (maximization) step: estimate class and item parameters – Iterate EM to maximize the likelihood of the parameters

15 Test against Data O = observed number of response patterns E = model estimated number of response patterns Pearson Chi-square based on likelihood ratio

16 Determine Number of Classes Substantive theory (parsimonious, interpretable) Predictive validity Auxiliary variables / covariates Statistical information and tests – Bayesian Information Criterion (BIC) – Entropy – Testing K against K-1 Classes Vuong-Lo-Mendell-Rubin likelihood-ratio test Bootstrapped likelihood ratio test

17 Bayesian Information Criterion (BIC) L = likelihood h = number of parameters N = sample size Choose model with smallest BIC BIC Difference > 4 appreciable

18 Quality of Classification

19 Testing K against K-1 Classes Bootstrapped likelihood ratio test LRT = 2[logL(model 1)- logL(model2)], where model 2 is nested in model 1. Bootstrap Steps: 1.Estimate LRT for both models 2.Use bootstrapped samples to obtain distributions for LRT of both models 3.Compare LRT and get p values

20 Testing K against K-1 Classes Vuong-Lo-Mendell-Rubin likelihood-ratio test

21 Determine Quality of Indicators Good indicators – Item response probability is close to 0 or 1 in each class Bad indicators – Item response probability is high in more than one classes, like cross-loading in factor analysis – Item response probability is low in all classes like low-loading in factor analysis

22 LCA Examples LCA LCA with covariates Class predicts a categorical outcome

23 Save Membership Variable Variable: idvar = id; Output: Savedata: File = cmmber.txt; Save = cprob;

24 Latent Profile Analysis

25 Finite Mixture Modeling ( multivariate normal variables) Finite = finite number of subgroups/classes Variables are normally distributed in each class Means differ across classes Variances are the same across Covariances can differ without restrictions or equal with restrictions across classes Latent profile can be special case with covariances fixed at zero.

26 Mixture Factor Analysis Allow one to examine measurement properties of items in heterogeneous subgroups / classes Measurement invariance is not required assuming heterogeneity Factor structure can change See Mplus outputs

27 Factor Mixture Analysis Parental Control Parental Acceptance Feel people in your family understand you Feel you want to leave home Feel you and your family have fun together Feel that your family pay attention to you Feel your parents care about you Feel close to your mother Feel close to your father Parents let you make your own decisions about the time you must be home on weekend nights Parents let you make your own decisions about the people you hang around with Parents let you make your own decisions about what you wear Parents let you make your own decisions about which television programs you watch Parents let you make your own decisions about what time you go to bed on week nights Parents let you make your own decisions about what you eat

28 Two dimensions of Parenting

29 Mixture SEM See mixture growth modeling

30 Mixture Modeling with Known Classes Identify hidden classes within known groups Under nonrandomized experiments – Impose equality constraints on covariates to identify similar classes from known groups – Compare classes that differ in covariates


Download ppt "Mixture Modeling Chongming Yang Research Support Center FHSS College."

Similar presentations


Ads by Google