170 likes | 609 Views
Temporal Causal Modeling with Graphical Granger Methods. SIGKDD 07 August 13, 2007. Andrew Arnold (Carnegie Mellon University) Yan Liu (IBM T.J. Watson Research) Naoki Abe (IBM T.J. Watson Research). Talk Outline. Introduction and motivation Overview of Granger causality
E N D
Temporal Causal Modeling with Graphical Granger Methods SIGKDD 07 August 13, 2007 Andrew Arnold (Carnegie Mellon University) Yan Liu (IBM T.J. Watson Research) Naoki Abe (IBM T.J. Watson Research)
Talk Outline • Introduction and motivation • Overview of Granger causality • Graphical Granger methods • Exhaustive Granger • Lasso Granger • SIN Granger • Vector auto-regression (VAR) • Experimental results
A Motivating Example: Key Performance Indicator Data (KPI)in Corporate Index Management [S&P] Variables Time
KPI Case Study: Temporal Causal Modeling for Identifying Levers of Corporate Performance • How can we leverage information in temporal data to assist causal modeling and inference ? • Key idea: A cause necessarily precedes its effects… Variables Time
Granger Causality • Granger causality • Introduced by the Nobel prize winning economist, Clive Granger [Granger ‘69] • Definition: a time series x is said to “Granger cause” another time series y, if and only if: • regressing for y in terms of past values of both y and x • is statistically significantly better than regressing y on past values of y only • Assumption: no common latent causes
Graphical Granger Methods • Exhaustive Granger • Test all possible univariate Granger models independently • Lasso Granger • Use L1-normed regression to choose sparse multivariate regression models • [Meinshausen & Buhlmann, ‘06] • SIN Granger • Do matrix inversion to find correlations between features across time • [Drton & Perlman, ‘04] • Vector auto-regression (VAR) • Fit data to linear-normal time series model • [Gilbert, ‘95]
Baseline methods: SIN and VAR • SIN • VAR
Empirical Evaluation of Competing Methods • Evaluation by simulation • Sample data from synthetic (linear normal) causal model • Learn using a number of competing methods • Compare learned graphs to original model • Measure similarity of output graph to original graph in terms of • Precision of predicted edges • Recall of predicted edges • F1 of predicted edges • Parameterize performance analysis • Randomly sample graphs from parameter space • Lag; Features; Affinity; Noise; Samples per feature; Samples per feature per lag • Conditioning to see interaction effects • E.g. Effect of # features when samples_per_feature_per_lag is small vs large
Experiment 1A: Performance vs. Factors- Random sampling all factors -
Experiment 1B: Performance vs. Factors- Fixing other factors - 13
Experiment 1C: Performance vs. Factors- Detail: Parametric Conditioning -
Experiment 3: Real World DataOutput Graphs on the Corporate KPI Data