Interspeech'2005 - Eurospeech

Lisbon, Portugal
September 4-8, 2005

Using Random Forest Language Models in the IBM RT-04 CTS System

Peng Xu (1), Lidia Mangu (2)

(1) Johns Hopkins University, USA; (2) IBM T.J. Watson Research Center, Yorktown Heights, NY, USA

One of the challenges in large vocabulary speech recognition is the availability of large amounts of data for training language models. In most state-of-the-art speech recognition systems, n-gram models with Kneser-Ney smoothing still prevail due to their simplicity and effectiveness. In this paper, we study the performance of a new language model, the random forest language model, in the IBM conversational telephony speech recognition system. We show that although the random forest language models are designed to deal with the data sparseness problem, they also achieve statistically significant improvements over n-gram models when the training data has over 500 million words.

Full Paper

Bibliographic reference.  Xu, Peng / Mangu, Lidia (2005): "Using random forest language models in the IBM RT-04 CTS system", In INTERSPEECH-2005, 741-744.