ISCA Archive Interspeech 2006
ISCA Archive Interspeech 2006

Forward-backwards training of hybrid HMM/BN acoustic models

Konstantin Markov, Satoshi Nakamura

In this paper, we describe an application of the Forward-Backwards (FB) algorithm for maximum likelihood training of hybrid HMM/Bayesian Network (BN) acoustic models. Previously, HMM/BN parameter estimation was based on a Viterbi training algorithm that requires two passes over the training data: one for BN learning and one for updating HMM transition probabilities. In this work, we first analyze the F-B training for a conventional HMM and show that the state PDF parameter estimation is analogous to weighted-data classifier training. The gamma variable of the Forward-Backwards algorithm plays the role of the data weight. From this perspective, it is straightforward to apply F-B-based training to the HMM/BN models since the BN learning algorithm allows training with weighted data. Experiments on accented speech (American, British and Australian English) show that F-B training outperforms the previous Viterbi learning approach and that the HMM/BN model achieved better performance than the conventional HMM.

doi: 10.21437/Interspeech.2006-224

Cite as: Markov, K., Nakamura, S. (2006) Forward-backwards training of hybrid HMM/BN acoustic models. Proc. Interspeech 2006, paper 1838-Mon3BuP.9, doi: 10.21437/Interspeech.2006-224

  author={Konstantin Markov and Satoshi Nakamura},
  title={{Forward-backwards training of hybrid HMM/BN acoustic models}},
  booktitle={Proc. Interspeech 2006},
  pages={paper 1838-Mon3BuP.9},