SamTaylor
- 20
- 0
Hi,
I try to teach myself Hidden Markov Models. I am using this text
"www.cs.sjsu.edu/~stamp/RUA/HMM.pdf" as Material. The Introduction with
the example was reasonable but now I have trouble in unterstanding
some of the derivation.
I can follow the math and use the formulas to get results but
I also want to understand the meaning behind it.
One question that arise at hidden markov models is to determine
the likelihood of the oberserved sequence O with the given Model
as written below:
b = probability of the observation
a = transiting from one state to another
\pi = starting probability
O = observation
X = state sequence
\lambda = hidden markov model
( Section 4, Page 6 in the Text )
P(O | X, \lambda) = b_{x0}(<img src="https://cdn.jsdelivr.net/joypixels/assets/8.0/png/unicode/64/1f635.png" class="smilie smilie--emoji" loading="lazy" width="64" height="64" alt="O_o" title="Er... what? O_o" data-smilie="12"data-shortname="O_o" />)* b_ {x1}(<img src="https://cdn.jsdelivr.net/joypixels/assets/8.0/png/unicode/64/1f635.png" class="smilie smilie--emoji" loading="lazy" width="64" height="64" alt="O_o" title="Er... what? O_o" data-smilie="12"data-shortname="O_o" />) *\cdots* b_{xT-1}(O_{T-1})
P(X|\lambda) = \pi_{x_0}a_{x_0,x_1}a_{x_1,x_2}*\cdots* a_{x_{T-1},x_{T-2}}
P(O, X|\lambda) = \frac{P(O \cap X \cap \lambda)}{P( \lambda)}
P(O | X, \lambda) *P(X|\lambda) = \frac{P(O \cap X \cap \lambda)}{P(X \cap \lambda)} \frac{P(X \cap \lambda)}{P(\lambda)}=\frac{P(O \cap X \cap \lambda)}{P(\lambda)}
P(O, X|\lambda) = P(O | X, \lambda) * P(X|\lambda)
P(O | \lambda) = \sum\limits_{X} P(O,X|\lambda)
P(O | \lambda) = \sum\limits_{X} P(O | X, \lambda) * P(X|\lambda)
My question is: Why do I get the likelihood of the oberserved sequence
by summing up over all all possible state sequences. Can someone please
explain it differently?
I try to teach myself Hidden Markov Models. I am using this text
"www.cs.sjsu.edu/~stamp/RUA/HMM.pdf" as Material. The Introduction with
the example was reasonable but now I have trouble in unterstanding
some of the derivation.
I can follow the math and use the formulas to get results but
I also want to understand the meaning behind it.
One question that arise at hidden markov models is to determine
the likelihood of the oberserved sequence O with the given Model
as written below:
b = probability of the observation
a = transiting from one state to another
\pi = starting probability
O = observation
X = state sequence
\lambda = hidden markov model
( Section 4, Page 6 in the Text )
P(O | X, \lambda) = b_{x0}(<img src="https://cdn.jsdelivr.net/joypixels/assets/8.0/png/unicode/64/1f635.png" class="smilie smilie--emoji" loading="lazy" width="64" height="64" alt="O_o" title="Er... what? O_o" data-smilie="12"data-shortname="O_o" />)* b_ {x1}(<img src="https://cdn.jsdelivr.net/joypixels/assets/8.0/png/unicode/64/1f635.png" class="smilie smilie--emoji" loading="lazy" width="64" height="64" alt="O_o" title="Er... what? O_o" data-smilie="12"data-shortname="O_o" />) *\cdots* b_{xT-1}(O_{T-1})
P(X|\lambda) = \pi_{x_0}a_{x_0,x_1}a_{x_1,x_2}*\cdots* a_{x_{T-1},x_{T-2}}
P(O, X|\lambda) = \frac{P(O \cap X \cap \lambda)}{P( \lambda)}
P(O | X, \lambda) *P(X|\lambda) = \frac{P(O \cap X \cap \lambda)}{P(X \cap \lambda)} \frac{P(X \cap \lambda)}{P(\lambda)}=\frac{P(O \cap X \cap \lambda)}{P(\lambda)}
P(O, X|\lambda) = P(O | X, \lambda) * P(X|\lambda)
P(O | \lambda) = \sum\limits_{X} P(O,X|\lambda)
P(O | \lambda) = \sum\limits_{X} P(O | X, \lambda) * P(X|\lambda)
My question is: Why do I get the likelihood of the oberserved sequence
by summing up over all all possible state sequences. Can someone please
explain it differently?
Last edited by a moderator: