Multi-Language Neural Network Language Models

Anton Ragni, Edgar Dakin, Xie Chen, Mark J.F. Gales, Kate M. Knill

In recent years there has been considerable interest in neural network based language models. These models typically consist of vocabulary dependent input and output layers and one, or more, hidden layers. A standard problem with these networks is that large quantities of training data are needed to robustly estimate the model parameters. This poses a challenge when only limited data is available for the target language. One way to address this issue is to make use of overlapping vocabularies between related languages. However this is only applicable to a small set of languages, and the impact is expected to be limited for more general applications. This paper describes a general solution that allows data from any language to be used. Here, only the input and output layers are vocabulary dependent whilst hidden layers are shared, language independent. This multi-task training set-up allows the quantity of data available to train the hidden layers to be increased. This multi-language network can be used in a range of configurations, including as initialisation for previously unseen languages. As a proof of concept this paper examines multilingual recurrent neural network language models. Experiments are conducted using language packs released within the IARPA Babel program.

DOI: 10.21437/Interspeech.2016-371

Cite as

Ragni, A., Dakin, E., Chen, X., Gales, M.J., Knill, K.M. (2016) Multi-Language Neural Network Language Models. Proc. Interspeech 2016, 3042-3046.

author={Anton Ragni and Edgar Dakin and Xie Chen and Mark J.F. Gales and Kate M. Knill},
title={Multi-Language Neural Network Language Models},
booktitle={Interspeech 2016},