Third Workshop on Spoken Language Technologies for Under-resourced Languages

Cape Town, South Africa
May 7-9, 2012

Multilingual Bottle-Neck Features and its Application for Under-Resourced Languages

Ngoc Thang Vu (1), Florian Metze (2), Tanja Schultz (1)

(1) Cognitive Systems Lab, Institute for Anthropomatics, Karlsruhe Institute of Technology (KIT), Germany
(2) Language Technologies Institute, Carnegie Mellon University (CMU), Pittsburgh, PA, USA

In this paper we present our latest investigation on multilingual bottle-neck (BN) features and its application to rapid language adaptation to new languages. We show that the overall performance of a Multilayer Perceptron (MLP) network improves significantly by initializing it with a multilingualMLP. Furthermore, ASR performance increases on both, on those languages which were used for multilingual MLP training, and on a new language. We propose a new strategy called “open target language” MLP to train more flexible models for language adaptation, which is particularly suited for small amounts of training data. The final results on the Vietnamese GlobalPhone database gave 15.8% relative improvement in terms of Syllable Error Rate (SyllER) for the ASR system trained with 22.5h data and 16.9% relative gains for the system trained with only 2h data.

Index Terms: multilingual bottle-neck feature, language adaptation

Full Paper

Bibliographic reference.  Vu, Ngoc Thang / Metze, Florian / Schultz, Tanja (2012): "Multilingual bottle-neck features and its application for under-resourced languages", In SLTU-2012, 90-93.