2016 ◽  
Vol 153 ◽  
pp. 29-36 ◽  
Author(s):  
Junseok Kwon ◽  
Ralf Dragon ◽  
Luc Van Gool

2000 ◽  
Vol 12 (4) ◽  
pp. 831-864 ◽  
Author(s):  
Zoubin Ghahramani ◽  
Geoffrey E. Hinton

We introduce a new statistical model for time series that iteratively segments data into regimes with approximately linear dynamics and learns the parameters of each of these linear regimes. This model combines and generalizes two of the most widely used stochastic time-series models—hidden Markov models and linear dynamical systems—and is closely related to models that are widely used in the control and econometrics literatures. It can also be derived by extending the mixture of experts neural network (Jacobs, Jordan, Nowlan, & Hinton, 1991) to its fully dynamical version, in which both expert and gating networks are recurrent. Inferring the posterior probabilities of the hidden states of this model is computationally intractable, and therefore the exact expectation maximization (EM) algorithm cannot be applied. However, we present a variational approximation that maximizes a lower bound on the log-likelihood and makes use of both the forward and backward recursions for hidden Markov models and the Kalman filter recursions for linear dynamical systems. We tested the algorithm on artificial data sets and a natural data set of respiration force from a patient with sleep apnea. The results suggest that variational approximations are a viable method for inference and learning in switching state-space models.


Sign in / Sign up

Export Citation Format

Share Document