2
$\begingroup$

Question: What is the stationary probability distribution $\pi(x)$ for the following Markov Chain?

Definition: An irreducible Markov chain with countably infinite state space is positive recurrent iff we can find a function $\pi: S \to [0,1]$ that satisfies

i.) $0 \leq \pi(x) \leq 1$ for all $x\in S$

ii.)\begin{equation*} \sum_{x \in S}\pi(x) = 1 \end{equation*} iii.)\begin{equation*} \pi(x) = \sum_{y\in S}\pi(y)p(y,x) \end{equation*} for all $x\in S$, where $p: S\times S \to [0,1]$ gives the transition probabilities for the Markov chain.

Markov Chain: Let $\{X_{n}\}_{n=0}^{\infty}$ be the Markov chain on state space $S=\mathbb{Z}^{+}\cup\{0\}=\{0,1,2,\ldots\}$ that is defined by transition probabilities \begin{equation*} \begin{cases} p(m,m+2) = p, & \text{for } m\in S \text{ with } m>0\\ p(m,m-1) = 1-p, & \text{for } m\in S \text{ with } m>0\\ p(0,2) = p, & \\ p(0,0) = 1-p, & \\ p(m_{1},m_{2}) = 0, & \text{for } m_{1},m_{2}\in S \text{ in all other cases}\\ \end{cases} \end{equation*} where $p\in(0,1)$ is a fixed parameter of the process.

We require $p\neq 0$ and $p\neq 1$ because in either of these extreme cases we would no longer be working with an irreducible Markov chain.

Source: Problem 2.4 from Introduction to Stochastic Processes, Second Edition by Gregory F. Lawler is determining when this Markov chain is transient by finding $\alpha(x)$, which I can do. This Markov chain is positive recurrent for $p<1/3$, transient for $p>1/3$, and null recurrent for $p=1/3$.

My problem comes when trying to calculate $\pi(x)$ in that I am convinced that I can find several functions $\pi(x)$ that satisfy the definition. If someone can show me why there is a unique solution for $\pi(x)$ in the specific case of $p=1/7$, which makes things nice, I would consider that explanation to be a satisfactory solution for my purposes.

2 Answers 2

2

The other answer gives an explicit way to find $\pi$, but I just want to reiterate the argument that specifically shows why such a $\pi$ is unique in this special case (since this was asked). Note that any stationary distribution satisfies $\pi(k) = (1-p)\pi(k+1) + p\pi(k-2)$ for $k \geq 2$ (simply by applying condition (iii) at $x=k$). So in other words the function $\pi$ satisfies the recursion $\pi(k+1) = \frac{1}{1-p} \pi(k) - \frac{p}{1-p} \pi(k-2)$, whenever $k \geq 2$. This implies that if $\pi$ exists, then $\pi$ is necessarily determined by $\pi(0), \pi(1),$ and $\pi(2)$. So it suffices to show that those three values are uniquely determined. Applying (iii) at $x=0$ gives $\pi(0) = (1-p)\pi(1)+(1-p)\pi(0)$, which shows that $\pi(1)$ is determined by $\pi(0)$. Applying (iii) at $x=1$ shows that $\pi(1) = (1-p)\pi(2)$, so that $\pi(2)$ is determined by $\pi(1)$ (and hence by $\pi(0)$). We have shown that $\pi(0)$ is therefore the only free parameter here. But the normalizing condition (ii) implies that scaling $\pi(0)$ must scale the entire sum (ii) and thus $\pi(0)$ must also be uniquely determined as well.

Using more advanced methods, you can prove the following

Theorem: If $X$ is an irreducible Markov Chain on a countable state space $S$, then stationary distributions exist iff the return time of some (and hence all) points has finite expectation. In this case the stationary distribution is unique.

The reason why irreducibility is needed here is because for a reducible chain, we can "split up" the chain into its recurrence classes, then consider invariant measures for the "restricted" chains associated to each of those classes, and even take convex combinations of those. So there can be uncountably many invariant measures for reducible chains, in general. The very extreme example of this is the constant chain $p(x,y)=1_{[x=y]}$.

  • 0
    Your insight that $\pi_{0}$ is the only free parameter in the system, which is fixed by the normalization, was exactly what I needed. As the other reply indicates, the cubic as 2 roots in (-1,1), which caused my confusion: I thought I had 2 free parameters: $\pi_{m} = c_{1}x_{1}^{m} + c_{2}x_{2}^{m}$ where $c_{1}$ and $c_{2}$ were both free. In fact, they can both be solved for in terms of $\pi_{0}$. The final answer for $\pi_{m}$ can be expressed elegantly in terms of the roots $x_{1}$ and $x_{2}$ that fall within $(-1,1)$.2017-02-26
2

Given the structure of this Markov chain, I'd do the following.

1) Check that the Markov chain is transitive: for all states $m_1$, $m_2$, there exists $n$ such that $p^n (m_1, m_2) > 0$. This guarantees that the function $\pi$, if it exists, is unique, and that the following computation does not depends on the starting point.

2) Choose a starting point. Here, $0$ seems a good choice. Put $\tilde{\pi} (0) := 1$.

3) Find a function $\tilde{\pi}$ which satisfies the functional equation. Here, you can do it step by step starting from $0$ (which is why this method is well adapted to this specific Markov chain):

$$1=\tilde{\pi} (0) = \sum_{y \in S} \tilde{\pi} (y) p(y,x) = \tilde{\pi} (0) p(0,0)+\tilde{\pi} (1) p(1,0) = (1-p)(1+\tilde{\pi} (1)),$$

so $\tilde{\pi} (1) = p/(1-p)$. Doing the same starting from $1$:

$$\frac{p}{1-p}=\tilde{\pi} (1) = (1-p)\tilde{\pi} (2),$$

so $\tilde{\pi} (2) = p/(1-p)^2$. For the other sites, we don't have the boundary effect from $0$, so we can proceed recursively. For $n \geq 2$:

$$\tilde{\pi} (n) = \sum_{y \in S} \tilde{\pi} (y) p(y,n) = \tilde{\pi} (n-2) p(n-2,n)+\tilde{\pi} (n+1) p(n+1,n) = p \tilde{\pi} (n-2) + (1-p)\tilde{\pi} (n+1),$$

which lets you deduce $\tilde{\pi} (n+1)$ from the lower orders. Thus the sequence $(\tilde{\pi} (n))$ satisfies a finite linear recurrence, which can be solved explicitly if you know the roots of the characteristic polynomial:

$$(1-p)X^3-X^2+p = (X-1)((1-p)X^2-pX-p)$$

Note that $p=1/3$ corresponds to the case where one of the two unknown roots is equal to $1$ (and thus its power do not decay to $0$). For $p<1/3$, the two unknown roots are in $(-1,1)$, and the sequence $(\tilde{\pi} (n))$ decays exponentially fast.

4) When you are done, the sequence $(\tilde{\pi} (n))$ you have computed gives you an invariant measure. In the null recurrent case, there isn't much more you can do. In the positive recurrent case, $\sum_{n \in S} \tilde{\pi} (n) <+\infty$, and you just have to put:

$$\pi (n) := \frac{\tilde{\pi} (n)}{\sum_{n \in S} \tilde{\pi} (n)}.$$

$\pi$s is non-negative, its sum is $1$, and since the functional equation is linear, it is satisfied by $\pi$.

  • 0
    +1, especially for the "find an invariant measure blindly and then normalize it" idea, which is quite useful.2017-02-23