220 likes | 230 Views
This paper proposes a statistical model for parsing and recognizing waveforms with significant shape variability. It introduces segmental Hidden Markov Models (HMMs) with random effects that allow for within-class shape variability. The model is evaluated on two datasets, turbulent flow measurements and ECG data, using cross-validation. The results demonstrate the effectiveness of the proposed approach in waveform analysis and recognition.
E N D
Segmental HiddenMarkov Models with Random Effects for Waveform Modeling Author: Seyoung Kim & Padhraic Smyth Presentor: Lu Ren
Outline • Introduction • Segmental HMMs • Segmental HMMs with Random Effects • Model Inference and Parameter Estimation • Experiment Results • Conclusion
Introduction Problem: Automatically parsing and recognizing waveforms based on their shapes. Example: analysis of waveforms from turbulent flow, discrimination an earthquakes in seismograph, waveform match and fault diagnosis in complex systems. Challenge: To address the problem of significant variability in wave shape is difficult for automated methods. Source of variability: shifts of the locations of prominent features, scaling along axes and measurement noise. The paper proposed a new statistical model that directly addresses within-class shape variability.
Standard discrete-time HMMs: generates noisy versions of piecewise constant shapes over time, since the observations within a sequence of states of the same value have constant mean. Segment HMMs: allows arbitrary distribution on run lengths and allow the mean to be a function of timewithin each segment but uses the same fixed parameters for all waveforms.
Segmental HMM with random effects: allow each group of observations have their own parameters that are still coupled together by an overall population prior. Specifically to say: allow the slopes and intercepts to vary according to a prior distribution
Segmental HMMs In a segmental HMM, Transition matrix In waveform modeling, is constrained to allow only left-to-right transitions and do not allow self-transitions. Duration distribution: (1) State k produces a segment of observations of length d from the segment distribution model. Segment model: (2) vector of regression coefficients for the intercept and slope design matrix and is the Gaussian noise.
The F-B algorithm for segmental HMMs takes into account the duration distribution, recursively computes (3) in the forward pass and (4) in the backward pass . The F-B algorithm scores a previously unseen waveform y by calculating the likelihood: (5)
Segmental HMMs with Random Effects Consider theth segment of length from the th individual waveform generated by state . (6) represents the mean regression parameters and represents the variation in regression parameters. and is independent of . Notice the model is equivalent to with . It can be shown the joint distribution of and are : (7) And the posterior distribution of can be written as: (8)
where (9) and (10)
Inference and parameters estimation Inference The likelihood of a waveform given fixed parameters, is: (11) The viterbi algorithm can compute the most likely state sequence. However, in F-B algorithm, needs to be calculated for all possible durations in each of the and at each recursion. Using direct inversion of the covariance matrix, , leads to an overall time complexity of . Based on Bayes’s rule, the likelihood of generated by state as (12)
By setting to as in equation (9), the likelihood can be reduced to: (13) Where In this way, the time complexities of the F-B and Viterbi algorithm are reduced to . Then EM algorithm is used to get the maximum-likelihood estimates of the parameters from a training set : Where the sate sequence implies segments in waveform .
Parameters estimation (14) Given the complete data, the log-likelihood decouples into four parts. Because only parts of the data are observed, EM is used to find the local maximum solution.
In E Step, the expected log likelihood of the complete data is (15) with respect to (16) In M step, the optimization problem decouples into four parts and closed form solutions exist for all of the parameters but in practice the algorithm often converges relatively slowly. Segmental HMM Random effect segmental HMM
Faster learning with ECME Expectation conditional maximization(ECM): Replace the M step of the EM with a sequence of Constrained or conditional maximizations (CM): The set of parameters is divided into subvectors and in the th CM step of the th iteration, is maximized over under the constrain of . ECME: some of the CM steps of the ECM are replaced by a maximization of the actual log likelihood instead of the expected complete data log likelihood. For random effects segmental HMMs, we partition the parameters Into and consider ECME with two CM steps as follows:
and get the update equations for Random effects segmental HMM for fluid-flow waveform data
Experimental Results • Two data sets were used: • Bubble-Probe Interaction Data (turbulent flow measurements): • 9-fold cross-validation with 5 waveforms in the training set and 43 waveforms in the test set. Another 72 waveforms were used as negative examples in the test. • 2. ECG Data (heartbeat cycles) • 4-fold cross-validation with 6 waveforms as a training set and 18 waveforms as the testing set. Another 22 abnormal hearbeats were used as negative examples in the test. Evaluation method: Average LogP score, Recognition Accuracy and Segmentation Quality (the mean squared difference between the observed data and regression model)
Segmentation result by segmental HMM Random effect segmental HMM ROC plot for bubble-probe data ROC plot for ECG data
Conclusions • Segmental HMMs including random effects allows an individual waveform to vary its shape in a constrained manner. • The ECME algorithm greatly improved the speed of convergence of parameter estimation compared to a standard EM approach.