This question was asked in a previous examination
Transfer from one state of a Markov chain to another is determined by getting a head or a tail tossing a coin. State j means: j heads were obtained. Initially the Markov chain is in state 0. Find transition probability matrix in 3 steps and average number
P is the (one-step) transition probability matrix of a Markov chain with state space {0, 1, 2, 3, 4 0.5 0.0 0.5 0.0 0.0 0.25 0.5 0.25 0.0 0.0 P=10.5 0.0 0.5 0.0 0.0 0.0 0.0 0.0 0.5 0.5 0.0 0.0 0.0 0.5 0.5/ (a) Draw a transition diagram. (b) Suppose the chain starts at time 0 in state 2. That is, Xo 2. Find E Xi (c)Suppose the chain starts at time 0 in any of the states with...