Decoders that make use of token-passing restrict their search space by various types of token pruning. With use of the Language Model Look-Ahead (LMLA) technique it is possible to increase the number of tokens that can be pruned without loss of decoding precision. Unfortunately, for token passing decoders that use single static pronunciation prefix trees, full n-gram LMLA increases the needed number of language model probability calculations considerably. In this paper a method for applying full n-gram LMLA in a decoder with a single static pronunciation tree is introduced. The experiments show that this method improves the speed of the decoder without an increase of search errors.
Bibliographic reference. Huijbregts, Marijn / Ordelman, Roeland / Jong, Franciska de (2008): "Fast n-gram language model look-ahead for decoders with static pronunciation prefix trees", In INTERSPEECH-2008, 1582-1585.