7 Random Processes: Discrete-time Markov Chains Flashcards
Evolutionary processes are random
We have looked @ where time is discrete (generations)
deterministic, maps and ODES
equations, given IC we could solve
but in reality chance is an important component of evolution
model using probabilistic tools
discrete vs continuous RVs
pmf
probability vector
probability density
cumulative functs
expected values
moments
discrete RV X with binomial distribution
parameter p
distribution function
generating function
expectation
variance
fⱼˣ
≡ Prob(X = j) =
(n)
(j) pʲ(1-p)ⁿ⁻ʲ
=Bin(n,j)
j=0,…,n
e.g. urn containing fraction p black balls fraction 1-p white balls
X is RV corresponding to drawing X=j black balls in n attempts (returning the balls)
(as you vary p small average (peak of distribution towards left) (p large peak towards right)
when p=0.5 distribution peak at RV x=10 when n=20)
generating funct: pgf
Gₓ(z) ≡ E(zˣ) ≡ Σⱼ₌₀ⁿ zʲ fⱼˣ = · · ·
= [pz + 1 − p]ⁿ
average:
differentiating pgf then setting z=1
E(X) = G′ₓ(z = 1) = np(pz + 1 − p)ⁿ−¹
=np
variance:
var(X) ≡ E(X²) − (E(X))²
= G′′ₓ(1) + G′ₓ(1) − [G′(1)]²
= np(1 − p)
stochastic process
A stochastic process is a collection of (discrete or continuous) random variables {Xₜ:t ∈ T}, where T is an index set and t ∈ T is here the discrete time index. The set of all
possible realizations (or range) of Xₜ
is S such that Xₜ ∈ S (see Sec. 7.5).
The sample sequence
{Xᵢ, . . . , Xⱼ} = {Xₜ ∈ S : t ∈ T˜
= [i, j] ⊂ T}
corresponds to the sample path
or stochastic realization of the process Xₜ on t ∈ T˜ = [i, j] which is a subset of T
e.g. Xₜ~ position each minute an organism moving one step left or right with equal probability
infinite line
over a period of 100 minutes, organism moves for total of 1000 minutes
This is an example of
a symmetric one-dimensional random walk,
the state space is S = Z and T˜ =
{0, 1, 2, . . . , 100} ⊂ T = {0, 1, 2, . . . , 1000}.
Typical sample paths, or stochastic realizations,
of this process for t ∈ T˜ .
{X_i,..,X_j}={X_t in S s.t t in [i,j]}
We notice that while S = Z, all paths
are here within {−20, . . . , −2, −1, 0, +1, +2, . . . , +20}.
There are relationships between the set of random variables {X_t}: they are not independent. The stochastic process X_t can be discrete or continuous in time and space, depending respectively on whether t is a discrete or continuous variable, and S is a discrete or continuous
set.
random walk
An example of discrete-time stochastic process is the one-dimensional “random walk” in which, at each discrete time increment ∆t random walker can take one discrete step to the right with prob. q or one step to left with prob. 1-q. The random walk is symmetric when q=1/2, and asymmetric otherwise.
position wrt origin
∆x
S= {. . . , −1, 0, 1, . . . }
position of random walker
probabilities
Position of the RWer is characterized by probability vector
p(t) = (. . . , p₋₁(t), p₀(t), p₁(t), . . .)ᵀ
where pⱼ(t) gives the probability for RWer to be @ position k after time t steps
t=0,1,…
and
Σ_{j∈S} pⱼ(t)=1
(probability conservation)
probability distribution
is the RW probability distribution
{p_{j∈S} (t)}
Discrete-time Markov chains (DTMC)
we want to know how the probability vector p(t) =
p_{i∈S} (t))
with
p_i∈S (t))=Prob {Xt = i ∈ S}
of X_t vary in time
Markov property
Markov process is s.t
Markov property, the process is
is “Markovian”, it is a “Markov process” (here a DTMC) and there is a systematic way to find p(t)
Consider DTMC and hence time is discrete
X_t has the Markov property if
Prob{Xₜ = iₜ|X₀ = i₀, . . . , Xₜ₋₁ = iₜ₋₁}
= Prob{Xₜ = iₜ|Xₜ₋₁ = iₜ₋₁}
iₖ ∈ S and k ∈ {0, 1, 2, . . . , t}
MEMORYLESS
MEMORYLESS
The Markov property therefore means that the value of Xₜ depends only of Xₜ₋₁ , its value at time t-1
=> a Markov process is often said to be memoryless
E.g DTMC
one step transition probability
1D random walk; each move is independent from the previous and the position at t depends on where the RWer is at t-1.
one step transition probability
pᵢⱼ(t)
one step transition probability
pᵢⱼ(t)
one step transition probability
pᵢⱼ(t)
the probability for X=i at time t+1 given that X=j at the previous time step t
when time is homogeneous
pᵢⱼ
one step transition matrix
one step transition matrix
P=(pᵢⱼ)
[p₁₁ p₁₂ p₁₃…]
[p₂₁ p₂₂ p₂₃…]
[p₃₁ p₃₂ p₃₃…]
all cols add to 1
with
pᵢᵢⱼ=1- sum_{i̸=j} pᵢⱼ
We assume that all are constant (time is homogeneous) and therefore , and we have
PROBABILITY CONSERVATION
sum_ i in S of pᵢⱼ=1
and pᵢⱼ≥ 0
stochastic matrix
one step transition matrix
Since all columns add up to 1, is a “stochastic matrix”.
P²,P³,… are also stochastic matrices
stochastic matrices are that all have eigenvalues , and at least
one eigenvalue is equal to 1.
l-step probability matrix
P⁽ˡ⁾
=Pˡ
gives the probability of transition from i to j in steps.
For DTMC, there is a simple relation between simple identity step probability matrices, or powers
of :Chapman-Kolmogorov equation (CKE):
Chapman-Kolmogorov equation (CKE):
Pˡ= Pˡ⁻ˢPˢ
for any ℓ > s > 0
=> Evolution of probability of probability vector
p(t)
=(p₀,p₁,…,pₙ)^T
of DTMC
X_t in S={0,1,…,n}
is obtained from the one-transition matrix using the CKE
p(t+1)=Pp(t)
gives
p(t)= Pᵗp(0)
{pᵢ(t)}ᵢ₌₀ᶦ⁼ⁿ is the probability distribution of X_tin S={0,1,…,n}
(note: pⱼ(t+1)= Σᵢ Pᵢⱼ pᵢ(t)
sum of prob. that X_t=i multiplied by prob. of one step transition i to j
Moments
From the probability distribution , or probability vector, we can obtain the moments of Xₜ
kth (“raw”) moment is the expected value of (Xₜ)ᵏ that is
E[( Xₜ)ᵏ] = Σ_{j∈S} jᵏ pⱼ(t)
Stationary probability distribution
In the long run
lim_t→∞ pᵢ(t)= πᵢ
and is the stationary probability distribution of
Xₜ∈ S = {0, 1, . . . }
The probability distribution π is stationary when it does not change in time
Pπ= π
HENCE π
is a right eigenvector of the one-step transition matrix P with eigenvalue 1
(In finite space S , the fact
that is a stochastic matrix P
of finite size ensures that there
is at least one such eigenvector.
However there can be more
than one.)
finite DTMC has a unique stat dist
A finite DTMC has unique stationary distribution π if X_t is irreducible (if it is always possible to reach any state from any other state) and aperiodic (if returns to any given state occur at irregular
intervals.)
Example from Q1(c) of Example Sheet 4:
X_t ∈ {1, 2, 3}
is a DTMC with 1-step transition matrix
P=(pᵢⱼ)
= Xₜ=1 Xₜ=2 Xₜ=3
Xₜ₊₁=1[ 0.5 0.5 0]
Xₜ₊₁=2[0.25 0.5 0.5]
Xₜ₊₁=3[0.25 0 0.5]
Stationary distribution obtained by solving the eigenvalue problem
Pπ= π =
(π₁)
(π₂)
(π₃)
solving gives
(2/5)
(2/5)
(1/5)
random walk
Random walk is broadly used to model movement in many biological applications, and in the continuum limits leads to the Brownian motion whose probability density obeys the diffusion equation.
In many applications, boundary conditions are important especially in the context of first-passage
problems.
Symmetric random walk on the infinite line Z
X_t
A random walker (RWer), initially at the origin (at t=0), moves along the infinite line.
At each time-step the RWer moves one step to the right with probability 1/2 or one step to the left also with probability 1/2
X_t: DTMC giving the position of the RWer after t time-steps (i.e. t jumps)
what is the probability that the RWer is at n∆x after k time steps?
We assume n and k of same parity and are interested in (both even/odd)
pₙ(k) = Prob{Xₖ= n∆x} = Probability{to be at n∆x after k time steps}
After k time steps: RWer is certainly between
between −k∆x and k∆x
(all k left, all k right)
Xₖ∈[-k∆x, k∆x]
At each time step, the RWer has two choices (left/right jump, L/R)
=> after t=1 time step, 2 possible paths with respect to the origin: either L or R
=> after t=2 time steps, 2x2 possible paths: L or R and again L or R => 4 possible paths: LL, LR, RL, RR
(-2,0,0,2)
=> after t=k , 2ᵏ possible paths: L or R at each step=> number of possible paths after k steps is 2ᵏ