1 / 10

Jiji Zhang and Ricardo Silva Lingnan University/University College London AISTATS 2011 – Fort Lauderdale, FL

Discussion of “Learning Equivalence Classes of Acyclic Models with Latent and Selection Variables from Multiple Datasets with Overlapping Variables”. Jiji Zhang and Ricardo Silva Lingnan University/University College London AISTATS 2011 – Fort Lauderdale, FL.

alia
Download Presentation

Jiji Zhang and Ricardo Silva Lingnan University/University College London AISTATS 2011 – Fort Lauderdale, FL

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Discussion of “Learning Equivalence Classes of Acyclic Models with Latent andSelection Variables from Multiple Datasets with Overlapping Variables” Jiji Zhang and Ricardo SilvaLingnan University/University College LondonAISTATS 2011 – Fort Lauderdale, FL

  2. On overlapping variables and partial information • Three main issues in statistical learning: estimation, computation and identification • The link to Spirtes (2001): “An anytime algorithm for causal inference” • The problem, then: estimating Markov equivalence classes when independence assessments stop at a particular order • T&S can be seen as a generalization in some directions: from incomplete independence assessments to more general equivalence classes

  3. Built-in robustness • Incorrect decisions on qualitative information might be less likely by pooling data from different sources “[It] has been suggested that causal discovery methods based solely on associations will find their greatest potential in longitudinal studies conducted under slightly varying conditions, where accidental independencies are destroyed and only structural independencies are preserved.” (Pearl, 2009, p.63)

  4. On selection bias • The assumption of common structure on separate studies also implies common selection bias • T&S make the role of structural assumptions very clear, but the one concerning common selection bias structure might not be as believable depending on the domain X1 X2 X3 X4 S X1 X2 X3 X4

  5. Combining different interventional studies • Structures might change given different interventional studies X1 X2 X3 X4 X1 X2 X3 X4 I X1 X2 X3 X4 I

  6. Beyond independence constraints • Cudeck (2000): overlapping factor analysis, finding 15 U U X3 X4 X3 X4 X5 X1 X2 X2

  7. Beyond independence constraints • Less obvious: what should be done with, e.g., additive error functional constraints X5 Dataset 1: X2 = f(X1, X5) + error X1 X2 X3 X4 Dataset 1I: X2 = g(X1) + error? X1 X2 X3 X4

  8. The Bayesian approach • Latent variable models might not be necessary, neither is the iid assumption. First likelihood function: L(1 | Dataset 1) X5 X1 X2 X3 X4 Second likelihood function: L(2 | Dataset 1I) X1 X2 X3 X4

  9. Related problems: finding substructure bygeneralizing penalized composite likelihood? • With/without same parameters, a single score function • S( | Data I, Data 2) = 1L(1 | Data1) + 2L(2| Data2) +  X5 First likelihood function: L( | Dataset 1) X1 X2 X3 X4 Second likelihood function: L( | Dataset 1I) X1 X2 X3 X4 X6

  10. Other approaches: generalizing penalized composite likelihood? • Enforcing constraints: link to constrained optimization Maximize S( | Data I, Data 2) = 1L(1 | Data1) + 2L(2 | Data2) + Subject to: for the corresponding independent models IM on each subset IM1(X1, X2, X3, X4)= IM2(X1, X2, X3, X4)

More Related