Stochastic Processes

Yuwei BaoSeptember 24, 2022

As I am taking Stochastic Processes class at Tulane University, I want to highlight some definitions from the textbook Essentials of Stochastic Processes (3rd Edition) by Rick Durrett (Springer, 2016)

Chapter 1 - Markov Chains

1.1 Definitions and Examples

Discrete type of Markov Chain

A sequence {Xn}\{X_n\} is a discrete Markov chain with transition matrix p(i,j)p(i,j) if for any j,i,in1j,i,i_{n-1}, s.t.

P(Xn+1=jXn=i,Xn1=in1,,X0=i0)=P(Xn+1=jXn=i)=p(i,j)P(X_{n+1} = j | X_n = i, X_{n-1} = i_{n-1}, \dots, X_0 = i_0) = P(X_{n+1} = j | X_n = i) = p(i,j)

1.2 Multistep Transition Probabilities

Theorem 1.1

The mm step transition probability P(Xn+m=jXn=1)P(X_{n+m} = j | X_n=1) is the mmth power of the transition matrix P:p×p×p××pP: p \times p \times p \times \dots \times p for mm times.

1.3 Classification of States

Chapman–Kolmogorov equation

pn+m(i,j)=kpm(i,k)pn(k,j)p^{n+m}(i,j) = \sum_{k} p^m(i,k) p^n (k,j)

pn+m(i,j)p^{n+m}(i,j) is the (n+m)(n+m) step transition probability.

Theorem 1.2 Strong Markov Property

Suppose TT is a stopping time. Given that T=nT=n and XT=yX_T =y, and other information about X0,X1,,XnX_0, X_1, \dots, X_n is irrelevant for predicting the future. Then XT+k,k0X_{T+k}, k \geq 0 behaves like a Markov Chain with initial state yy.

Lemma 1.3

Suppose Px(Tyk)α>0P_x(T_y \leq k) \geq \alpha > 0 for all xx in the state space SS for a Markov Chain XnX_n. Then

Px(Ty>nk)(1α)nP_x(T_y > nk) \leq (1-\alpha)^n

Lemma 1.4

If xyx \rightarrow y and yzy \rightarrow z, then xzx \rightarrow z

Theorem 1.5

If ρxy>0\rho_{xy} >0 but ρyx<1\rho_{yx} <1, then xx is a transit state.

Lemma 1.6

If xx is recurrent and ρxy>0\rho_{xy} > 0, then ρxy=1\rho_{xy} = 1.

Theorem 1.7

If CC is a finite closed and irreducible set, then all states in CC are recurrent.

Theorem 1.8

If the state space S of a Markov Chain is finite, then S can be written as the disjoint union TR1RkT \cup R_1 \cup \dots \cup R_k, where T is the set of transient states and Ri,1ikR_i, 1 \leq i \leq k are closed irreducible sets of recurrent states.

Lemma 1.9

If xx is recurrent and xyx \rightarrow y, then yy is recurrent.

Lemma 1.10

In a finite closed set, there must be at least one recurrent state.

Lemma 1.11

ExN(y)=ρxy1ρyyE_x N(y) = \frac{\rho_{xy}}{1- \rho_{yy}}

where ExN(y)E_x N(y) denotes the expected number of visits to state yy beginning at state xx.

Lemma 1.12

ExN(y)=n=1pn(x,y)E_x N(y) = \sum_{n=1}^\infty p^n(x,y)

Theorem 1.13

A state yy is recurrent if and onl if

k=1pn(y,y)=EyN(y)=\sum_{k=1}^\infty p^n(y,y) = E_y N(y) = \infty

1.4 Stationary Distributions

Stationary Distributions

Suppose we start a Markov Chain in a randomly selected state. \ We have

P(Xn=j)=iP(X0=i,Xn=j)=iP(X0=i)P(Xn=jX0=i)=iq(i)pn(i,j), where q(t)=p(X0=i)\begin{align*} P(X_n = j) &= \sum_i P(X_0 = i, X_n = j) \\ &= \sum_i P(X_0 = i) P(X_n = j| X_0 = i) \\ &= \sum_i q(i) p^n(i,j), \text{ where } q(t) = p(X_0 = i) \end{align*}

If the chain has kk states, then

q=(q1q2qk)q = \begin{pmatrix} q_1 \\ q_2 \\ \vdots \\ q_k \end{pmatrix}

is a kk-dimensional vector, and the transition matrix pp is k×kk \times k

1.4.1 Doubly Stochastic Chains

Definition 1.2 Doubly Stochastic Chains

A transition matrix PP is doubly stochastic if its columns sum to 1. xp(x,y)=1\sum_x p(x,y) = 1

Theorem 1.14

If PP is a doubly stochastic transition matrix for a Markov chain over NN states, then the uniform distribution π(x)=1N,x=1,,N\pi(x) = \frac{1}{N}, x = 1, \dots, N is a stationary distribution.

1.5 Detailed Balance Condition

Detailed balance

A stationary distribution π\pi satisfies the detailed \textbf{balance condition} if π(x)p(x,y)=π(y)p(y,x)\pi(x) p(x,y) = \pi(y) p(y,x)

1.5.1 Reversibility

Theorem 1.15

Fix nn and let Ym=XnmY_m = X_{n-m} for 0mn0 \leq m \leq n. Then YmY_m is a Markov chain with transition probability

1.5.2 The Metropolis-Hastings Algorithm

The Metropolis-Hastings Algorithm

Consider sampling from some distribution π\pi. If we begin in state XX, we will jump to another state with probability

r(x,y)=min{π(y)q(y,x)π(x)q(x,y),1}\begin{align*} r(x,y) = \min \{ \frac{\pi(y)q(y,x)}{\pi(x) q(x,y)}, 1\} \end{align*}

where q(x,y)q(x,y) is a jumping distribution The transition probability from the chain is thus

p(x,y)=q(x,y)r(x,y)p(x,y) = q(x,y) r(x,y)

Suppose that π(y)q(x,y)>π(x)q(x,y)\pi(y) q(x,y) > \pi(x) q(x,y), then r(x,y)=1r(x,y) = 1, so π(x)p(x,y)=π(x)q(x,y)1=π(x)q(x,y)\pi(x) p(x,y) = \pi(x) q(x,y) 1 = \pi(x) q(x,y) and π(y)p(y,y)=π(y)q(y,x)r(y,x)=π(y)q(y,x)π(x)q(x,y)π(y)q(y,x)=π(x)q(x,y)\pi(y) p(y,y) = \pi(y) q(y,x ) r(y,x) = \pi(y) q(y,x) \frac{\pi(x)q(x,y)}{\pi(y) q(y,x)} = \pi(x) q(x,y), so we see that the detailed balance condition is satisfied.

To generate a sample, we can run the chain for a sufficient period of time to reach an equilibrium.

1.5.3 Kolmogerov Cycle Condition

Kolmogerov Cycle Condition

Consider an irreducible Markov chain with state space SS. We say that for an irreducible Markov Chain with state space ss, the cycle condition is satisfied if, given a cycle of states x0,x1,,xn1,xnx_0, x_1, \dots, x_{n-1}, x_n with p(xi1,xi)>01inp(x_{i-1},x_i) >0 \quad 1 \leq i \leq n, we have

i=1np(xi1,xi)=i=1np(xi,xi1)\prod_{i=1}^n p(x_{i-1},x_i) = \prod_{i=1}^n p(x_i,x_{i-1})

Theorem 1.16

There is a stationary distribution that satisfies the detailed balance condition if and only if the cycle condition holds.

1.6 Limit Behavior

Limit behavior

If yy is a transient state, then we know that n=1pn(x,y)<\sum_{n=1}^{\infty} p^n(x,y) < \infty for any state xx \textbf{Lemma 1.11} so limnpn(x,y)=0\lim_{n \rightarrow \infty} p^n(x,y) = 0. So we can focus on studying recurrent states.

Lemma 1.17

If ρxy>0\rho_{xy}>0 and ρyx>0\rho_{yx}>0, the xx and yy have the same period.

Lemma 1.18

If p(x,x)>0p(x,x)>0, then xx has period 1.

4 key assumptions about a chain with transition matrix $P$

  • II: PP is irreducible
  • AA: All states have period 1 (aperiodic)
  • RR: All states are recurrent
  • SS: There is a stationary distribution π\pi

Theorem 1.19 Convergence Theorem

Suppose I,A,SI,A,S hold, then as nn \rightarrow \infty, pn(x,y)π(y)p^n(x,y) \rightarrow \pi(y)

Theorem 1.20 Asymptotic Frequency

Suppose I,RI,R hold, let Nn(y)N_n(y) be the number of visits to yy up to time nn, then Nn(y)n1EyTy\frac{N_n(y)}{n} \rightarrow \frac{1}{E_y T_y}

Theorem 1.21

If II and SS hold, then π(y)=1EyTy\pi(y) = \frac{1}{E_y T_y}, and hence the stationary distribution is unique.

Theorem 1.22

Suppose I,SI,S hold, and xf(x)π(x)<\sum_x |f(x)| \cdot \pi(x) < \infty for some f(x)f(x), then

1nm=1nf(xm)xxf(x)π(x)\frac{1}{n} \sum_{m=1}^n f(x_m) \rightarrow^x \sum_x f(x) \pi(x)

Theorem 1.23 Convergence Theorem

Suppose I,SI,S hold. Then

1nm=1npm(x,y)π(y)\frac{1}{n} \sum_{m=1}^n p^m(x,y) \rightarrow \pi(y)

Theorem 1.24

Suppose pp is irreducible and recurrent. Let xSx \in S and let Tx=inf{n1:Xn=x}T_x = \inf \{n \geq 1: X_n = x\}

μx(y)=n=0Px(Xn)=y,Tx>n)\mu_x(y) = \sum_{n=0}^{\infty} P_x (X_n) = y, T_x > n)

defines a stationary measure with 0<μx(y)<0 < \mu_x(y) < \infty for all yy.

1.8 Proof of the Convergence Theorem

Lemma 1.25

If there is a stationary distribution, then all states yy that have π(y)>0\pi(y)>0 are recurrent.

Lemma 1.26

If xx has period 1, then there is a number n0n_0 s.t. if nn0n \geq n_0, then nIxn \in I_x (the set of all times kk for which pk(x,x)>0p^k(x,x) >0)

Lemma 1.27

IxI_x is closed under addition. That is, if i,jIxi,j \in I_x, then i+jIxi+j \in I_x

1.9 Exit Distributions

Theorem 1.28

Consider a Markov chain with state space SS. Let AA and BB be subsets of SS, so that C=S(AB)C=S-(A \cup B) is finite. Suppose h(a)=1h(a)=1 for aA,h(b)=0a \in A, h(b) = 0 for bBb \in B, and that for xCx \in C we have

h(x)=yp(x,y)h(y)h(x) = \sum_y p(x,y) h(y)

If Px(VAVB<)>0xCP_x(V_A \wedge V_B < \infty) > 0 \quad \forall x \in C, then h(x)=Px(Va<Vb)h(x) = P_x(V_a < V_b).

1.10 Exit Times

Theorem 1.29

Let VA=inf{n0;XnA}V_A = \inf \{ n \geq 0; X_n \in A\}. Suppose C=SAC=S-A is finite, and that Px(VA<)>0xCP_x(V_A < \infty) >0 \quad \forall x \in C. If g(a)=0aAg(a) = 0 \quad \forall a \in A, and xC\forall x \in C, we have

g(x)=1+yp(x,y)g(y)g(x) = 1 + \sum_y p(x,y)g(y)

then g(x)=Ex(VA)g(x) = E_x (V_A)

1.11 Infinite State Spaces

Warning

xx is positive recurrent if ExTx<E_xT_x < \infty

If a state is recurrent but not positive recurrent, i.e. Px(Tx<)=1P_x(T_x < \infty) = 1 but ExTx=E_x T_x = \infty, then we say that xx is null recurrent.

Theorem 1.30

For an irreducible chain the following are equivalent:

  1. Some state is positive recurrent.
  2. There is a stationary distribution π\pi.
  3. All states are positive recurrent.