Presentation is loading. Please wait.

Presentation is loading. Please wait.

Assessing Estimability of Latent Class Models Using a Bayesian Estimation Approach Elizabeth S. Garrett Scott L. Zeger Johns Hopkins University Departments.

Similar presentations


Presentation on theme: "Assessing Estimability of Latent Class Models Using a Bayesian Estimation Approach Elizabeth S. Garrett Scott L. Zeger Johns Hopkins University Departments."— Presentation transcript:

1 Assessing Estimability of Latent Class Models Using a Bayesian Estimation Approach Elizabeth S. Garrett Scott L. Zeger Johns Hopkins University Departments of Oncology and Biostatistics ENAR March 26, 2001 Charlotte, NC

2 Motivation: We would like to investigate how many classes of depression exist using a latent class model applied to an epidemiologic sample of mental health symptoms. Issue: Latent class models require a “large” sample size to be estimable. Questions: Is our sample size large enough? How many classes can we fit “reliably” given the amount of data that we have?

3 Latent Class Model Overview (binary indicator case) There are M classes of depression.  m represents the proportion of individuals in the population in class m (m = 1,…,M). Each individual is a member of one of the M classes, but we do not know which. The latent class of individual i is denoted by  i (  i  1,…M). Symptom prevalences vary by class. The prevalence for symptom j in class m is denoted by p mj. Given class membership, the symptoms are independent.

4 Likelihood Function For individual i,

5 Two Scenarios A latent class model with M classes may be only weakly estimable if (1) M classes are too many (over-parameterization). “True” number of classes is < M. (2) There are “truly” M classes, but there is not enough data to identify all of the classes. Better to use fewer classes than to use the M class model with weakly estimated parameters.

6 Estimation of LC Models How large is large enough? Depends on –true class sizes –prevalence of items/symptoms –measurement error in items/symptoms Example 1: Three class model N = 500. True class sizes are 30%, 40%, 30%. Symptom prevalences range from 30% to 80%. Example 2: Three class model N = 500. True class sizes are 80%, 18%, 2%. Symptom prevalences range from 3% to 10%. Same sample size, but are both estimable?

7 Distinction and Definitions Statistical identifiability refers to the “difficulty of distinguishing among two or more explanations of the same empirical phenomena.” 1 Not so much a concept that stems from limited data, but the ability to distinguish between explanations even if we had unlimited data. Classic example: –X is random variable such that E(X) =  1 -  2 –observations of X allow us to identify  1 -  2 –but, we cannot identify  1 or  2 regardless of the amount of information we collect on X. –Infinitely many values of  1 and  2 can give rise to the same value of X 1 Franklin Fisher, “Statistical Identifiability”, International Encyclopedia of Statistics, ed. Kruskal and Tanur, 1977.

8 “Identifiability” in Latent Class Models Latent class issues with identifiability are “twofold”. Still have the statistical identifiability to contend with. –Local identifiability issues have been well-defined –Goodman, 1974, Biometrika –McHugh, 1956, Biometrika –Bandeen-Roche et al., 1997, JASA To distinguish issues stemming from limited data from classical identifiability we use alternative terminology: estimability. We say that a model is estimable in this context if there is enough data to uniquely estimate all parameters with some degree of certainty.

9 Local Identifiability Define p 0 to be a vector of parameters which defines the LC model. If, for all y and p in the neighborhood of p 0, then we say that L is locally identifiable. Maximum likelihood estimation is concerned with local identifiability NOT locally identified!

10 Weak Identifiability / Weak Estimability Bayesian concept Assume we partition parameter vector  = (  1,  2 ). If, then we say that  2 is not estimable. The data (Y) provides no information about  2 given  1. If most of the information we have about  2 is supplied by its prior distribution then: In words, if the prior distribution of  2 is approximately the same as its posterior, then we say that  2 is only weakly estimable or weakly identifiable.

11 Estimation Approaches Maximum Likelihood Approach: Find estimates of p, , and  that are most consistent with the data that we observe conditional on number of classes, M. Often used: EM algorithm (iterative fitting procedure) Bayesian Approach : Quantify beliefs about p, , and  before and after observing data. Often used: Gibbs sampler, MCMC algorithm

12 Bayesian Model Review Every parameter, , in a Bayesian model has a prior distribution associated with it: The resulting posterior distribution is a combination of the the prior distribution and the the likelihood function: If there is a lot of information in the data, then the likelihood will provide most of the information that determines the posterior. If there is not a lot of information in the data, then the prior will provide most of the information that determines the posterior

13 Estimability Assessment Approach Compare posterior distribution to the prior distribution –visual inspection: picture tells us how much information is coming from the prior and how much from the likelihood. –quantify similarity/difference by a statistic (e.g. percent overlap). Bayes factor is also related to this idea.

14 Simulated Data Examples

15 LCED for 2 class data

16 Simulated Data Examples

17 LCED for 3 class data

18 Latent Class Analysis in Mental Health Research

19 Epidemiologic Catchment Area (ECA) Study Goal: To obtain epidemiologic sample on mental health data. Population: Community dwelling individuals over age 18 in five sites in the US. Instrument: Diagnostic Interview Schedule (DIS) includes 17 depression symptoms. Our sample: –Data from 1981, Baltimore site only –Full information on depression symptoms as defined in the DSM-III on 2938 individuals.

20 Depression Symptom Groups

21 LCED for ECA data

22 Conclusions from ECA example The four class model is not an appropriate model for this data. This could be due to (1) Four class model is an overparameterization (2) More than three classes are needed to describe depression, but the data set is too small to estimate more classes.

23 Aside: Looking at  2 statistics May say just compare  2 statistics from a maximum likelihood estimation procedure BUT!  2 relies on assumption that most patterns are relatively prevalent which is not generally true. May see significant differences due to summing up very small differences over large number of samples. M and M -1 class models are not really “nested” in interpretation Ô Not a valid statistic for comparing LC models. See Bayes Factor and BIC: these are better statistics for comparing LC models.

24 Extension: Quantifying Estimability Calculate percent overlap between posterior and prior. Estimate “estimabilility index” for each class where there are K symptoms: Characteristics: –0 < < 1 –Larger numbers indicate weak estimability Also see: Bayes Factor


Download ppt "Assessing Estimability of Latent Class Models Using a Bayesian Estimation Approach Elizabeth S. Garrett Scott L. Zeger Johns Hopkins University Departments."

Similar presentations


Ads by Google