1 / 17

Latent Class Regression Model Graphical Diagnostics Using an MCMC Estimation Procedure

Latent Class Regression Model Graphical Diagnostics Using an MCMC Estimation Procedure. Elizabeth S. Garrett Scott L. Zeger Johns Hopkins University esg@jhu.edu. Overview. Latent class models can be useful tools for measuring latent constructs.

vine
Download Presentation

Latent Class Regression Model Graphical Diagnostics Using an MCMC Estimation Procedure

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Latent Class Regression Model Graphical Diagnostics Using an MCMC Estimation Procedure Elizabeth S. Garrett Scott L. Zeger Johns Hopkins University esg@jhu.edu

  2. Overview • Latent class models can be useful tools for measuring latent constructs. • Latent class model checking is somewhat complicated because we cannot “check” model fit using standard approaches which rely on comparing fitted values to observed. • After fitting a latent class regression model, what can we do to see if we obey several key assumptions? • Conditional independence? • Non-differential measurement?

  3. N=1126 in 1993 in Baltimore Symptoms (DSM-IV): dysphoria weight/appetite change sleep problems slow/increased movement loss of interest/pleasure fatigue guilt concentration problems thoughts of death Covariates of interest gender age marital status education income How are education and income associated with depression? From standard LC model fit: The symptoms listed at left define depression Depression is a latent class variable with 3 classes Classes are “ordered”: None Mild Severe What is the association between depression and socio-economic status?Epidemiologic Catchment Area (ECA) Study

  4. There are J classes of individuals. pj represents the proportion of individuals in the population in class j (j=1,…,J) Each person is a member of one of the J classes, but we do not know which. The latent class of individual i is denoted by ci. Symptom prevalences vary by class. The prevalence for symptom m in class j is denoted by mj. We assume that covariates, x, are associated with class membership Given class membership, the symptoms are independent of each other Given class membership, the symptoms are independent of covariates Latent Class Regression Model: Main Ideas CONDITIONAL INDEPENDENCE NON-DIFFERENTIAL MEASUREMENT

  5. Latent Class Regression Likelihood Assumptions where • Conditional Independence: • given an individual’s depression class, his/her symptoms are independent • P(yig, yih | ci) = P(yig | ci) P(yih | ci) • Non-differential Measurement: • given an individual’s depression class, covariates are not associated with symptoms • P(yig | xi, ci) = P(yig | ci)

  6. Latent Class Regression Results

  7. Depression Example:LCR coefficients (log ORs) * indicates significant at the 0.10 level Note: class 1 is non-depressed,class 2 is mild, class 3 is severe

  8. Checking Conditional Independence Assumption • For each pair of symptoms (h and g), in each class (j), consider • If assumption holds, this OR will be approximately equal to 1. • (The log OR will be approximately equal to 0). • Why may this get tricky? • We don’t KNOW class assignments. • Need a strategy for assigning individuals to classes. Checking Non-differential Measurement Assumption • For each symptom (h), covariate (x), and class (j) combination, we can • estimate an odds ratio. Example in the binary covariate case:

  9. Model Estimation: Markov Chain Monte Carlo procedure • Bayesian Approach • Quantify beliefs about p, , and c before and after observing data. • Prior Probability: What we believe about unknown parameters before observing data. • Posterior Probability: What we believe about the parameters after observing data. • Model specifications: • Specify prior probability distribution: P(p, , c) • Combine prior with likelihood to obtain posterior distribution: P(p, , c|Y) P(p, , c) x L(Y| p, , c) • Estimate posterior distribution for each parameter using iterative procedure. P(p1|Y) =  P(p, , |Y)

  10. Bayesian Estimation Approach The Gibbs Sampler is an iterative process used to estimate posterior distributions of parameters. • we sample parameters from conditional distributions e.g. P(p1|Y, p, c, ) • At each iteration, we get ‘sampled’ values of p, , and c. • We use the samples from the iterations to estimate posterior distributions by averaging over other parameter values. This is a key feature for these methods!

  11. Checking Assumptions: MCMC (Bayesian) approach • At each iteration in the Gibbs sampler, individuals are automatically assigned to classes no need to “manually” assign. • At each iteration, simply calculate the log OR’s of interest. • Then, “marginalize” or average over all iterations. • Result is posterior distribution of log OR • From posterior distribution, we have both a point estimate and precisionestimate of the log OR. • We can calculate “posterior intervals” (similar to confidence intervals) to see if there is evidence of violation of assumptions.

  12. Checking Conditional Independence

  13. Checking Conditional Independence

  14. Checking Non-Differential Measurement

  15. Checking Non-Differential Measurement

  16. Implementation • “Canned” implementation: • BUGS (unix and linux) • WinBugs (windows) • http://astor.som.jhmi.edu/~esg/software • Scripts can be (have been) written in • R, Splus • SAS

  17. Checking Assumptions: Maximum Likelihood Approach Using ML approach, we can get a result that will likely be quite similar • (a) assign individuals to “pseudo-classes” based on posterior probability of class membership • (b) calculate OR’s within classes. • (c) repeat (a) and (b) at least a few times • (d) compare OR’s to 1. Drawback: • In ML, additional post hoc computations are necessary. • Don’t get precision estimates as you do in MCMC approach. • MCMC approach is designed for computing posterior distribution of functions of parameters.

More Related