scholarly journals A Probability Inequality for the Occupation Measure of a Reversible Markov Chain

1995 ◽  
Vol 5 (1) ◽  
pp. 37-43 ◽  
Author(s):  
I. H. Dinwoodie
2021 ◽  
Vol 58 (2) ◽  
pp. 372-393
Author(s):  
H. M. Jansen

AbstractOur aim is to find sufficient conditions for weak convergence of stochastic integrals with respect to the state occupation measure of a Markov chain. First, we study properties of the state indicator function and the state occupation measure of a Markov chain. In particular, we establish weak convergence of the state occupation measure under a scaling of the generator matrix. Then, relying on the connection between the state occupation measure and the Dynkin martingale, we provide sufficient conditions for weak convergence of stochastic integrals with respect to the state occupation measure. We apply our results to derive diffusion limits for the Markov-modulated Erlang loss model and the regime-switching Cox–Ingersoll–Ross process.


2017 ◽  
Vol 114 (11) ◽  
pp. 2860-2864 ◽  
Author(s):  
Maria Chikina ◽  
Alan Frieze ◽  
Wesley Pegden

We present a statistical test to detect that a presented state of a reversible Markov chain was not chosen from a stationary distribution. In particular, given a value function for the states of the Markov chain, we would like to show rigorously that the presented state is an outlier with respect to the values, by establishing a p value under the null hypothesis that it was chosen from a stationary distribution of the chain. A simple heuristic used in practice is to sample ranks of states from long random trajectories on the Markov chain and compare these with the rank of the presented state; if the presented state is a 0.1% outlier compared with the sampled ranks (its rank is in the bottom 0.1% of sampled ranks), then this observation should correspond to a p value of 0.001. This significance is not rigorous, however, without good bounds on the mixing time of the Markov chain. Our test is the following: Given the presented state in the Markov chain, take a random walk from the presented state for any number of steps. We prove that observing that the presented state is an ε-outlier on the walk is significant at p=2ε under the null hypothesis that the state was chosen from a stationary distribution. We assume nothing about the Markov chain beyond reversibility and show that significance at p≈ε is best possible in general. We illustrate the use of our test with a potential application to the rigorous detection of gerrymandering in Congressional districting.


1999 ◽  
Vol 13 (1) ◽  
pp. 95-101 ◽  
Author(s):  
Olivier François

We give a new bound on the spectral gap of an ergodic time reversible Markov chain in term of the conductance of the chain.


Sign in / Sign up

Export Citation Format

Share Document