Hi,
I am implementing a speech recognition system based on Hidden Markov models. I
have finished the Baum-Welch algorithm for model training - and tested it on
some primitive models.
I work with discrete HMM, so I am using a code book with 50 output symbols. For
speech parametrization I am using 12 MFCC coefficients for every block of
sound(512 sound samples - 30 ms).
Now when I am training the model on several speech samples (10 times the same
word is recorded). I am getting the emission probability of some of the symbols
from code book equal ZERO in all model states (after training). Thant
wouldn't be a problem. But sometimes during recognition process I get a
symbol, which has for this particular model probability of emission in all
states equal 0. (Even when it is the same word). That's why, the total
probabilty for this word to be generated by this model is equal ZERO too. (and
that is obviously wrong)
Is there any posibility to avoid this situation? (Maybe more sound samples
during model training - to cover all the posibilities how the word can be spoken
by the speaker). I haven't seen any note about this problem in articles
about HMM and speech recognition.
I hope I explained my situation clear enough.
Thanks in advance.
Hidden Markov model - zero probability problem
Started by ●December 7, 2010