Sixth European Conference on Speech Communication and Technology
This paper address the problem of training HMMs using long files of uninterrupted speech with limited and constant memory requirements. The classical training algorithms usually require limited duration training utterances due to memory constraints for storing the generated trellis. Our solution allows to exploits databases that are transcribed, but not partitioned into sentences, using a sliding window Forward-Backward algorithm. This approach has been tested on the connected digits TI/NIST database and on long sequences of Italian digits. Our experimental results show that for a lookahead value L of about 1-2 sec it is possible to achieve reestimation counts that are affected by errors less than 1.e-7, producing similar reestimated models. Another application of our sliding window Forward-Backward algorithm is MMIE training, that we have tested on the TI/NIST database connected digits using as a general model the recognition tree rather than the N-best hypotheses, or the word lattices.
Full Paper (PDF) Gnu-Zipped Postscript
Bibliographic reference. Chesta, C. / Laface, Pietro / Nigra, M. (1999): "Piecewise HMM discriminative training", In EUROSPEECH'99, 2729-2732.