250 likes | 296 Views
Explore the application of hierarchical Dirichlet process prior for modeling nonlinear phenomena using switching dynamical processes. Utilize a Gibbs sampler for efficient learning from various datasets.
E N D
Laboratory for Information and Decision Systems Nonparametric Bayesian Learning of Switching Dynamical Processes Emily Fox, Erik Sudderth, Michael Jordan, and Alan Willsky Nonparametric Bayes Workshop 2008 Helsinki, Finland
= set of dynamic parameters Priors on Modes • Switching linear dynamical processes useful for describing nonlinear phenomena • Goal: allow uncertainty in number of dynamical modes • Utilize hierarchical Dirichlet process (HDP) prior • Cluster based on dynamics Switching Dynamical Processes
Outline • Background • Switching dynamical processes: SLDS, VAR • Prior on dynamic parameters • Sticky HDP-HMM • HDP-AR-HMM and HDP-SLDS • Sampling Techniques • Results • Synthetic • IBOVESPA stock index • Dancing honey bee
Vector autoregressive (VAR) process: Linear Dynamical Systems • State space LTI model:
State space models VAR processes Linear Dynamical Systems • State space LTI model: • Vector autoregressive (VAR) process:
Switching VAR process: Switching Dynamical Systems • Switching linear dynamical system (SLDS):
Group all observations assigned to mode k Define the following mode-specific matrices Place matrix-normal inverse Wishart prior on: Prior on Dynamic Parameters Rewrite VAR process in matrix form: Results in K decoupled linear regression problems
Sticky HDP-HMM Infinite HMM: Beal, et.al., NIPS 2002HDP-HMM: Teh, et. al., JASA 2006Sticky HDP-HMM: Fox, et.al., ICML 2008 • Dirichlet process (DP): • Mode space of unbounded size • Model complexity adapts to observations • Hierarchical: • Ties mode transition distributions • Shared sparsity • Sticky: self-transition bias parameter Time Mode
Mode-specific transition distributions: sparsity of b is shared,increased probability of self-transition Sticky HDP-HMM • Global transition distribution:
HDP-SLDS HDP-AR-HMM and HDP-SLDS HDP-AR-HMM
Blocked Gibbs Sampler Sample parameters • Approximate HDP: • Truncate stick-breaking • Weak limit approximation: • Sample transition distributions: • Sample dynamic parameters using state sequence as VAR(1) pseudo-observations: Fox, et.al., ICML 2008
Blocked Gibbs Sampler Sample mode sequence • Use state sequence as pseudo-observations of an HMM • Compute backwards messages: • Block sample as:
Blocked Gibbs Sampler Sample state sequence • Equivalent to LDS with time-varying dynamic parameters • Compute backwards messages (backwards information filter): • Block sample as: All Gaussian distributions
Hyperparameters • Place priors on hyperparameters and learn them from data • Weakly informative priors • All results use the same settings hyperparameters can be set using the data
HDP-VAR(1)-HMM HDP-VAR(2)-HMM HDP-HMM HDP-SLDS Results: Synthetic VAR(1) 5-mode VAR(1) data
HDP-VAR(1)-HMM HDP-VAR(2)-HMM HDP-HMM HDP-SLDS Results: Synthetic AR(2) 3-mode AR(2) data
HDP-VAR(1)-HMM HDP-VAR(2)-HMM HDP-HMM HDP-SLDS Results: Synthetic SLDS 3-mode SLDS data
sticky HDP-SLDS non-sticky HDP-SLDS ROC Results: IBOVESPA Daily Returns • Data: Sao Paolo stock index • Goal: detect changes in volatility • Compare inferred change-points to 10 cited world events Carvalho and Lopes, Comp. Stat. & Data Anal., 2006
x-pos y-pos • Observation vector: • Head angle (cosq, sinq) • x-y body position sinq cosq Results: Dancing Honey Bee • 6 bee dance sequences with expert labeled dances: • Turn right (green) • Waggle (red) • Turn left (blue) Sequence 1 Sequence 2 Sequence 3 Sequence 4 Sequence 5 Sequence 6 Time Oh et. al., IJCV, 2007
Nonparametric approach: Model: HDP-VAR(1)-HMM Set hyperparameters Unsupervised training from each sequence Infer: Number of modes Dynamic parameters Mode sequence Supervised Approach [Oh:07]: Model: SLDS Set number of modes to 3 Leave one out training: fixed label sequences on 5 of 6 sequences Data-driven MCMC Use learned cues (e.g., head angle) to propose mode sequences Results: Dancing Honey Bee Oh et. al., IJCV, 2007
HDP-AR-HMM: 83.2%SLDS [Oh]: 93.4% HDP-AR-HMM: 93.2%SLDS [Oh]: 90.2% HDP-AR-HMM: 88.7%SLDS [Oh]: 90.4% Results: Dancing Honey Bee Sequence 4 Sequence 5 Sequence 6
HDP-AR-HMM: 46.5%SLDS [Oh]: 74.0% HDP-AR-HMM: 44.1%SLDS [Oh]: 86.1% HDP-AR-HMM: 45.6%SLDS [Oh]: 81.3% Results: Dancing Honey Bee Sequence 1 Sequence 2 Sequence 3
Conclusion • Examined HDP as a prior for nonparametric Bayesian learning of SLDS and switching VAR processes. • Presented efficient Gibbs sampler • Demonstrated utility on simulated and real datasets