Transition Matrix in Markov Chain Explained
Markov chain is a very important math model. Many practical problems can be modeled as Markov chain, and then be solved. I would like to talk about the transition matrix, specifically how to deduce the transition matrix after step n. Let us use a simple but non-trivial example. Let say we have only two states: State 1 and State 2. State 1 has 0.6 probability to stay at itself, and 0.4 probability to go to State 2. State 2 has 0.8 probability to stay at itself and 0.2 probability to go to State 1. We can easily to write down the transition matrix below: $$q = \begin{pmatrix} 0.6 & 0.4 \\ 0.2 & 0.8 \\ \end{pmatrix} $$ Please notice, in different literature, people tend to use different way of representing transition matrix. In our example, each row represents the corresponding probability for each state. Let us assume we start at $T_0$, which has PMF of $(1,0)$. It simply means at $T_0$, it is 100% at State 1. What happen after 1 step at $T_1$, what will the PMF cha...