Download presentation
Presentation is loading. Please wait.
Published byHelen McGee Modified over 9 years ago
1
Segmental Hidden Markov Models with Random Effects for Waveform Modeling Author: Seyoung Kim & Padhraic Smyth Presentor: Lu Ren
2
Outline Introduction Segmental HMMs Segmental HMMs with Random Effects Model Inference and Parameter Estimation Experiment Results Conclusion
3
Introduction Problem: Automatically parsing and recognizing waveforms based on their shapes. Example: analysis of waveforms from turbulent flow, discrimination an earthquakes in seismograph, waveform match and fault diagnosis in complex systems. Challenge: To address the problem of significant variability in wave shape is difficult for automated methods. Source of variability: shifts of the locations of prominent features, scaling along axes and measurement noise. The paper proposed a new statistical model that directly addresses within-class shape variability.
5
Standard discrete-time HMMs: generates noisy versions of piecewise constant shapes over time, since the observations within a sequence of states of the same value have constant mean. Segment HMMs: allows arbitrary distribution on run lengths and allow the mean to be a function of time within each segment but uses the same fixed parameters for all waveforms.
6
Segmental HMM with random effects: allow each group of observations have their own parameters that are still coupled together by an overall population prior. Specifically to say: allow the slopes and intercepts to vary according to a prior distribution
7
Segmental HMMs In a segmental HMM, Transition matrix In waveform modeling,is constrained to allow only left-to-right transitions and do not allow self-transitions. Duration distribution: (1) State k produces a segment of observations of length d from the segment distribution model. Segment model: (2) vector of regression coefficients for the intercept and slope design matrix and is the Gaussian noise.
8
The F-B algorithm for segmental HMMs takes into account the duration distribution, recursively computes in the forward pass and in the backward pass. The F-B algorithm scores a previously unseen waveform y by calculating the likelihood: (3) (4) (5)
9
Segmental HMMs with Random Effects Consider the th segment of length from the th individual waveform generated by state. represents the mean regression parameters and represents the variation in regression parameters. and is independent of. Notice the model is equivalent to with. It can be shown the joint distribution of and are : And the posterior distribution ofcan be written as: (6) (7) (8)
10
where and (9) (10)
11
Inference and parameters estimation The likelihood of a waveformgiven fixed parameters, is: The viterbi algorithm can compute the most likely state sequence. However, in F-B algorithm, needs to be calculated for all possible durations in each of the and at each recursion. Using direct inversion of the covariance matrix,, leads to an overall time complexity of. Based on Bayes’s rule, the likelihood of generated by state as (11) (12) Inference
12
By setting to as in equation (9), the likelihood can be reduced to: (13) Where In this way, the time complexities of the F-B and Viterbi algorithm are reduced to. Then EM algorithm is used to get the maximum-likelihood estimates of the parameters from a training set : Where the sate sequence implies segments in waveform.
13
Given the complete data, the log-likelihood decouples into four parts. Because only parts of the data are observed, EM is used to find the local maximum solution. (14) Parameters estimation
14
In E Step, the expected log likelihood of the complete data is with respect to (15) (16) In M step, the optimization problem decouples into four parts and closed form solutions exist for all of the parameters but in practice the algorithm often converges relatively slowly. Segmental HMMRandom effect segmental HMM
15
Faster learning with ECME Expectation conditional maximization(ECM): Replace the M step of the EM with a sequence of Constrained or conditional maximizations (CM): The set of parameters is divided into subvectors and in the ECME: some of the CM steps of the ECM are replaced by a maximization of the actual log likelihood instead of the expected complete data log likelihood. th CM step of theth iteration, is maximized over under the constrain of. For random effects segmental HMMs, we partition the parameters Into and consider ECME with two CM steps as follows:
16
and get the update equations for Random effects segmental HMM for fluid-flow waveform data
18
Experimental Results Two data sets were used: 1.Bubble-Probe Interaction Data (turbulent flow measurements): 9-fold cross-validation with 5 waveforms in the training set and 43 waveforms in the test set. Another 72 waveforms were used as negative examples in the test. 2. ECG Data (heartbeat cycles) 4-fold cross-validation with 6 waveforms as a training set and 18 waveforms as the testing set. Another 22 abnormal hearbeats were used as negative examples in the test. Evaluation method: Average LogP score, Recognition Accuracy and Segmentation Quality (the mean squared difference between the observed data and regression model)
20
Segmentation result by segmental HMMRandom effect segmental HMM ROC plot for ECG data ROC plot for bubble-probe data
21
Conclusions 1.Segmental HMMs including random effects allows an individual waveform to vary its shape in a constrained manner. 2. The ECME algorithm greatly improved the speed of convergence of parameter estimation compared to a standard EM approach.
22
Thanks!
Similar presentations
© 2024 SlidePlayer.com Inc.
All rights reserved.