1. Properties
$Proof.$ First, we prove the “if” statement. ($⇐$)
- $X_0 \sim \lambda$ (Markov Chain definition)
- Next, we start from the definition of the conditional probability of $X_n$ given the history and prove the Markov property for $(X_t)_{t⩾0}$
This proves that, if $\forall x_0, x_1,\cdots, x_n \in I:\mathbb{P}\left(X_{0}=x_{0}, X_{1}=x_{1}, \cdots, X_{n}=x_{n}\right) = \displaystyle \lambda_{x_{0}} \cdot \prod_{i=1}^n p_{x_{i-1} x_{i}}$, then $(X_t)_{t⩾0} \sim\text{Markov}(\lambda, P)$.
Now, we prove the “only if” statement ($⇒$). We begin with the definition of conditional probability followed by the Markov property.
We recursively use Markov property. Specifically, in the above expression we try to rewrite $\mathbb{P}\left(X_{0}=x_{0}, X_{1}=x_{1}, \ldots, X_{n-1}=x_{n-1}\right) $ as a conditional probability followed by use of Markov property.
We next consider some consequences of Markov property.
$\textbf{Remark 5.3.}$ As a consequence of the Markov property ($\textbf{Theorem 5.1}$), we can show that if $(X_t)_{t⩾0}\sim \text{Markov}(\lambda, P)$, then given $X_k= x$, we have $(X_t)_{t⩾k}\sim \text{Markov}(\delta_x, P)$.
This remark reinforces the idea that Markov chains do not have memory. If we consider the state of a Markov chain at any arbitrary time instant, the random process starting that instant is also Markov. It has the same transition probability matrix; however, the initial distribution is degenerate.
$\textbf{Lemma 5.4.}$ If $P$ and $Q$ are transition matrices on $I$, then $P Q$ is a transition matrix on $I$.
$Proof.$ The entries of product matrix $P Q$ can be written as
The sum of all entries in a row is computed as the sum of $(P Q)_{ij}$ overall $j$, as
Since every entry in the $P Q$ matrix is non-negative and the row sum of the matrix is 1, we know that $P Q$ is a transition probability matrix (row stochastic).
$\textbf{Corollary 5.5.}$ If $P$ is a transition probability matrix, then $P$ n is also a transition probability matrix.
$Proof.$ We can prove this statement easily by induction. The base case is when $n = 1$. This is true trivially. We prove, given that the statement is true for $n = k−1$, it is also true for $n = k$. Begin by rewriting $P^k$ as a product of $P$ and $P^{k−1}$ . From our assumption, $P^{k−1}$ is a transition probability matrix. We know from $\textbf{Lemma 5.4}$ that since both $P$ and $P^{k−1}$ are transition probability matrix, $P^k$ is also a transition matrix.
Note that the converse is not true. $P^n$ being a transition matrix is insufficient for $P$ to be a transition matrix. A counter-example can be easily constructed.
$\textbf{Counter-example}\ (n = 2)$: Consider a transition matrix $A^{2}=\left[\begin{array}{ll}1 & 0 \\ 0 & 1\end{array}\right]$. We can write $A^2 = AA$ where $A=\left[\begin{array}{cc}1 / \sqrt{2} & 1 / \sqrt{2} \\ 1 / \sqrt{2} & -1 / \sqrt{2}\end{array}\right]$. This counter-example proves that $A^2$ being a transition matrix is not sufficient for $A$ to be a transition matrix.
$\textbf{Remark 5.6.}$ If $(X_t)_{t⩾0} \sim \text{Markov}(\lambda, P)$, then
- $\mathbb{P}\left(X_{n}=i\right)=\left(\lambda P^{n}\right)_{i}$
- $\mathbb{P}\left(X_{n}=j \mid X_{0}=i\right)=\left(P^{n}\right)_{i j}$
$Proof.$
- We know from $\textbf{Theorem 5.1}$, $\displaystyle \mathbb{P}\left(X_{0}=x_{0}, X_{1}=x_{1}, \cdots, X_{n}=x_{n}\right)=\lambda_{x_{0}} \cdot \prod_{i=1}^n p_{x_{i-1} x_{i}}$ . We can write the probability of a state being $i$ as,
- We begin by using the definition of conditional probability and $\textbf{Theorem 5.1.}$ Moreover, we use $\text{Fubini’s theorem}$ recursively.
$\textbf{Corollary 5.7.}$ Let $(X_t)_{t⩾0} \sim \text{Markov}(\lambda, P)$. Then,
2. Hitting Time
Intuitively, $H_i^A$ 指的是在初始状态为 $i$ 的情况下达到状态 $A$ 所需要的时间
$\textbf{Remark 5.9.}$ Let us look at hitting time of $A$ for some special cases:
- $H^A = 0$ if $X_0 \in A$
- $H^A = 1$ if $X_0 \not \in A$ and $X_1 \in A$
- $H^A = n$ if $X_0, X_1, .\ldots , X_{n−1} \not \in A$ and $X_n \in A$
Note that the hitting time of a set is a random variable that takes values in $\{0, 1, 2, \ldots\} \cup \{∞\}$
$\textbf{Remark 5.10.}$ We denote the probability that the hitting time of set $A$ is finite as $h^A$. That is
We can also consider probability of finite hitting time conditioned on the initial state,
Clearly, using law of total probability.
$\textbf{Corollary 5.11.}$ If $h_i^A < 1$, then, $\mathbb E[H_i^A ] = ∞$
The proof for this corollary is simple. One can write down the definition of expectation and note that since $h_i^A < 1$ for some $i$, hitting time when starting at $i$ is infinity $H_i^A = ∞$ with non-zero probability. Hence the expectation of hitting time is infinity. Consider the probability that $H_i^A$ is finite, $h_i^A = \mathbb P_i(H^A < ∞)$. The following result is true.