INTERSPEECH 2009
10th Annual Conference of the International Speech Communication Association

Brighton, United Kingdom
September 6-10, 2009

Improving Speech Understanding Accuracy with Limited Training Data Using Multiple Language Models and Multiple Understanding Models

Masaki Katsumaru (1), Mikio Nakano (2), Kazunori Komatani (1), Kotaro Funakoshi (2), Tetsuya Ogata (1), Hiroshi G. Okuno (1)

(1) Kyoto University, Japan
(2) Honda Research Institute Japan Co. Ltd., Japan

We aim to improve a speech understanding module with a small amount of training data. A speech understanding module uses a language model (LM) and a language understanding model (LUM). A lot of training data are needed to improve the models. Such data collection is, however, difficult in an actual process of development. We therefore design and develop a new framework that uses multiple LMs and LUMs to improve speech understanding accuracy under various amounts of training data. Even if the amount of available training data is small, each LM and each LUM can deal well with different types of utterances and more utterances are understood by using multiple LM and LUM. As one implementation of the framework, we develop a method for selecting the most appropriate speech understanding result from several candidates. The selection is based on probabilities of correctness calculated by logistic regressions. We evaluate our framework with various amounts of training data.

Full Paper

Bibliographic reference.  Katsumaru, Masaki / Nakano, Mikio / Komatani, Kazunori / Funakoshi, Kotaro / Ogata, Tetsuya / Okuno, Hiroshi G. (2009): "Improving speech understanding accuracy with limited training data using multiple language models and multiple understanding models", In INTERSPEECH-2009, 2735-2738.