Question: Consider a Markov chain {X_n: n = 0,1,....} with the following transition probability matrix P={0 0.45 0.55 0.5 0.2 0.3 0.1 0.9 1 } Assuming
Consider a Markov chain {X_n: n = 0,1,....} with the following transition probability matrix
P={0 0.45 0.55
0.5 0.2 0.3
0.1 0.9 1 }
Assuming the chain is three times as likely to being at time=0 in states 2 or 3, find E(X_2)
Step by Step Solution
There are 3 Steps involved in it
1 Expert Approved Answer
Step: 1 Unlock
Question Has Been Solved by an Expert!
Get step-by-step solutions from verified subject matter experts
Step: 2 Unlock
Step: 3 Unlock
