Finding probability from a markov chain

Mathematics
Tutor: None Selected Time limit: 1 Day

this chain has a transition matrix of:

.4 .6

.8 .2

If, on the first observation, the system is in state 2, what state is the system most likely to occupy on the third observation? (If there is more than one such state, which is the first one.) 

Apr 20th, 2015

2nd:  P(s1) = 0.8 P(S2) = 0.2 

3nd P(S1) = 0.4*0.8 +0.8*0.2 = 0.32+0.16 = 0.48

P(S2) = 0.6*0.8+0.2*0.2 = 0.52

so it is state 2 

Apr 20th, 2015

Studypool's Notebank makes it easy to buy and sell old notes, study guides, reviews, etc.
Click to visit
The Notebank
...
Apr 20th, 2015
...
Apr 20th, 2015
Dec 9th, 2016
check_circle
Mark as Final Answer
check_circle
Unmark as Final Answer
check_circle
Final Answer

Secure Information

Content will be erased after question is completed.

check_circle
Final Answer