9th Annual Conference of the International Speech Communication Association

Brisbane, Australia
September 22-26, 2008

Fast N-Gram Language Model Look-Ahead for Decoders with Static Pronunciation Prefix Trees

Marijn Huijbregts, Roeland Ordelman, Franciska de Jong

University of Twente, The Netherlands

Decoders that make use of token-passing restrict their search space by various types of token pruning. With use of the Language Model Look-Ahead (LMLA) technique it is possible to increase the number of tokens that can be pruned without loss of decoding precision. Unfortunately, for token passing decoders that use single static pronunciation prefix trees, full n-gram LMLA increases the needed number of language model probability calculations considerably. In this paper a method for applying full n-gram LMLA in a decoder with a single static pronunciation tree is introduced. The experiments show that this method improves the speed of the decoder without an increase of search errors.

Full Paper

Bibliographic reference.  Huijbregts, Marijn / Ordelman, Roeland / Jong, Franciska de (2008): "Fast n-gram language model look-ahead for decoders with static pronunciation prefix trees", In INTERSPEECH-2008, 1582-1585.