1 / 34

CCD Summer Course Day2 Afternoon

Causal inference

Udayan
Download Presentation

CCD Summer Course Day2 Afternoon

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Center for Causal Discovery Day 2: Search June 9, 2015 Carnegie Mellon University 1

  2. Outline Day 2: Search 1.  Bridge Principles: Causation ßà Probability 2.  D-separation 3.  Model Equivalence 4.  Search Basics (PC, GES) 5.  Latent Variable Model Search (FCI) 6.  Examples 2

  3. Tetrad Demo and Hands On 3

  4. Tetrad Demo and Hands-on 1)  Go to “estimation2” 2)  Add Search node (from Data1) - Choose and execute one of the “Pattern searches” 3)  Add a “Graph Manipulation” node to search result: “choose Dag in Pattern” 4)  Add a PM to GraphManip 5)  Estimate the PM on the data 6)  Compare model-fit to model fit for true model 4

  5. Backround Knowledge Tetrad Demo and Hands-on 1)  Create new session 2)  Select “Search from Simulated Data” from Template menu 3)  Build graph below, PM, IM, and generate sample data N=1,000. 4)  Execute PC search, α = .05 5

  6. Backround Knowledge Tetrad Demo and Hands-on 1)  Add “Knowledge” node – as below 2)  Create “Required edge X3 à X1 as shown below. 3)  Execute PC search again, α = .05 4)  Compare results (Search2) to previous search (Search1) 6

  7. Backround Knowledge Tetrad Demo and Hands-on 1)  Add new “Knowledge” node 2)  Create “Tiers” as shown below. 3)  Execute PC search again, α = .05 4)  Compare results (Search2) to previous search (Search1) 7

  8. Backround Knowledge Direct and Indirect Consequences True Graph PC Output PC Output 8 Background Knowledge No Background Knowledge

  9. Backround Knowledge Direct and Indirect Consequences True Graph Direct Consequence Of Background Knowledge PC Output PC Output Indirect Consequence 9 Background Knowledge No Background Knowledge Of Background Knowledge

  10. Charitable Giving (Search) 1)  Load in charity data 2)  Add search node 3)  Enter Background Knowledge: •  Tangibility is exogenous •  Amount Donated is endogenous only •  Tangibility à Imaginability is required 4)  Choose and execute one of the “Pattern searches” 5)  Add a “Graph Manipulation” node to search result: “choose Dag in Pattern” 6)  Add a PM to GraphManip 7)  Estimate the PM on the data 8)  Compare model-fit to hypothetical model 10

  11. Constraint-based Search for Patterns 1) Adjacency phase 2) Orientation phase 11

  12. Constraint-based Search for Patterns: Adjacency phase X and Y are not adjacent if they are independent conditional on any subset that doesn’t X and Y 1) Adjacency •  Begin with a fully connected undirected graph •  Remove adjacency X-Y if X _||_ Y | any set S

  13. Causal Graph Independcies X1 X2 X1 X4 {X3} X1 X3 X4 X2 X4 {X3} X2 X1 Begin with: X3 X4 X2 From X1 X3 X1 X2 X4 X2 From X1 X4 {X3} X1 X3 X4 X2 From X1 X4 {X3} X2 X3 X4 X2

  14. Constraint-based Search for Patterns: Orientation phase 2) Orientation •  Collider test: Find triples X – Y – Z, orient according to whether the set that separated X-Z contains Y •  Away from collider test: Find triples X à Y – Z, orient Y – Z connection via collider test •  Repeat until no further orientations •  Apply Meek Rules 14

  15. Search: Orientation Patterns Test: X _||_ Z | S, is Y ∈S Y Unshielded Z X Y No Yes Non-Collider Collider Z X Y Z X Y Z X Y Z X Y Z X Y

  16. Search: Orientation Away from Collider Test Conditions 1) X1 - X2 adjacent, and into X2. 2) X2 - X3 adjacent 3) X1 - X3 not adjacent X3 X1 X2 Test X1 _||_ X3 | S, X2 ∈ ∈ S No Yes X1 X1 X3 X3 X2 X2

  17. Search: Orientation X1 After Adjacency Phase X3 X4 X2 Collider Test: X1 – X3 – X2 X1 X3 X4 X1 _||_ X2 X2 Away from Collider Test: X1 X1 à X3 – X4 X2 à X3 – X4 X3 X4 X1 _||_ X4 | X3 X2 X2 _||_ X4 | X3

  18. Away from Collider Power! X2 X3 X1 _||_ X3 | S, X2 ∈ ∈ S X1 X2 X1 X3 X2 – X3 oriented as X2à X3 Why does this test also show that X2 and X3 are not confounded? C X2 X1 X3 X2 X1 X3 X1 _||_ X3 | S, X2 ∈ ∈ S, C∉ ∉ S X1 _||_ X3 | S, X2 ∈ ∈ S

  19. Independence Equivalence Classes: Patterns & PAGs • Patterns (Verma and Pearl, 1990): graphical representation of d-separation equivalence among models with no latent common causes • PAGs: (Richardson 1994) graphical representation of a d- separation equivalence class that includes models with latent common causes and sample selection bias that are d-separation equivalent over a set of measured variables X 19

  20. PAGs: Partial Ancestral Graphs X1 X2 PAG X3 Represents X1 X2 X1 X2 T1 X3 X3 etc. X1 X2 X1 X2 T1 T1 T2 X3 X3 20

  21. PAGs: Partial Ancestral Graphs Z1 Z2 PAG X Y Represents Z1 Z2 Z1 Z2 T1 X3 X3 Y Y etc. Z1 Z2 Z2 Z1 T1 T2 T1 X3 X3 Y Y 21

  22. PAGs: Partial Ancestral Graphs What PAG edges mean. X1 and X2 are not adjacent X1 X2 X2 is not an ancestor of X1 X1 X2 No set d-separates X2 and X1 X1 X2 X1 is a cause of X2 X1 X2 X1 X2 There is a latent common cause of X1 and X2 22

  23. PAG Search: Orientation PAGs Y Unshielded Z X Y X _||_ Z | Y X _||_ Z | Y Non-Collider Collider Z X Y Z X Y

  24. PAG Search: Orientation X1 After Adjacency Phase X3 X4 X2 Collider Test: X1 – X3 – X2 X1 X3 X4 X1 _||_ X2 X2 Away from Collider Test: X1 X1 à X3 – X4 X2 à X3 – X4 X3 X4 X1 _||_ X4 | X3 X2 X2 _||_ X4 | X3

  25. Interesting Cases X1 X2 L L M1 M2 Z Y1 Y2 Y X L1 L L1 Z1 L2 X Y Z1 X Z2 Z2 M3 Y M4 25

  26. Tetrad Demo and Hands-on 1)  Create new session 2)  Select “Search from Simulated Data” from Template menu 3)  Build graphs for M1, M2, M3 “interesting cases”, parameterize, instantiate, and generate sample data N=1,000. L 4)  Execute PC search, α = .05 M1 5)  Execute FCI search, α = .05 Z Y X X1 X2 L L M3 Z1 Y1 Y2 M2 X Y Z2 L1 26

  27. Regression & Causal Inference 27

  28. Regression & Causal Inference Typical (non-experimental) strategy: 1.  Establish a prima facie case (X associated with Y) Z But, omitted variable bias X Y 2.  So, identifiy and measure potential confounders Z: a)  prior to X, b)  associated with X, c)  associated with Y 3. Statistically adjust for Z (multiple regression) 28

  29. Regression & Causal Inference Multiple regression or any similar strategy is provably unreliable for causal inference regarding X à Y, with covariates Z, unless: •  X prior to Y •  X, Z, and Y are causally sufficient (no confounding) 29

  30. Tetrad Demo and Hands-on 1)  Create new session 2)  Select “Search from Simulated Data” from Template menu 3)  Build a graph for M4 “interesting cases”, parameterize as SEM, instantiate, and generate sample data N=1,000. 4)  Execute PC search, α = .05 5)  Execute FCI search, α = .05 30

  31. Summary of Search 31

  32. Causal Search from Passive Observation •  PC, GES àPatterns (Markov equivalence class - no latent confounding) •  FCI àPAGs (Markov equivalence - including confounders and selection bias) •  CCD à Linear cyclic models (no confounding) •  BPC, FOFC, FTFC à (Equivalence class of linear latent variable models) •  Lingam à unique DAG (no confounding – linear non-Gaussian – faithfulness not needed) •  LVLingam à set of DAGs (confounders allowed) •  CyclicLingam à set of DGs (cyclic models, no confounding) •  Non-linear additive noise models à unique DAG •  Most of these algorithms are pointwise consistent – uniform consistent algorithms require stronger assumptions 32

  33. Causal Search from Manipulations/Interventions What sorts of manipulation/interventions have been studied? •  Do(X=x) : replace P(X | parents(X)) with P(X=x) = 1.0 •  Randomize(X): (replace P(X | parents(X)) with PM(X), e.g., uniform) •  Soft interventions (replace P(X | parents(X)) with PM(X | parents(X), I), PM(I)) •  Simultaneous interventions (reduces the number of experiments required to be guaranteed to find the truth with an independence oracle from N-1 to 2 log(N) •  Sequential interventions •  Sequential, conditional interventions Time sensitive interventions •  33

  34. Tetrad Demo and Hands-on 1)  Search for models of Charitable Giving 2)  Search for models of Foreign Investment 3)  Search for models of Lead and IQ 34

More Related