Analysis of Multilingual Sequence-to-Sequence Speech Recognition Systems

Martin Karafiát, Murali Karthick Baskar, Shinji Watanabe, Takaaki Hori, Matthew Wiesner, Jan Černocký


This paper investigates the applications of various multilingual approaches developed in conventional deep neural network - hidden Markov model (DNN-HMM) systems to sequence-to-sequence (seq2seq) automatic speech recognition (ASR). We employ a joint connectionist temporal classification-attention network as our base model. Our main contribution is separated into two parts. First, we investigate the effectiveness of the seq2seq model with stacked multilingual bottle-neck features obtained from a conventional DNN-HMM system on the Babel multilingual speech corpus. Second, we investigate the effectiveness of transfer learning from a pre-trained multilingual seq2seq model with and without the target language included in the original multilingual training data. In this experiment, we also explore various architectures and training strategies of the multilingual seq2seq model by making use of knowledge obtained in the DNN-HMM based transfer-learning. Although both approaches significantly improved the performance from a monolingual seq2seq baseline, interestingly, we found the multilingual bottle-neck features to be superior to multilingual models with transfer learning. This finding suggests that we can efficiently combine the benefits of the DNN-HMM system with the seq2seq system through multilingual bottle-neck feature techniques.


 DOI: 10.21437/Interspeech.2019-2355

Cite as: Karafiát, M., Baskar, M.K., Watanabe, S., Hori, T., Wiesner, M., Černocký, J. (2019) Analysis of Multilingual Sequence-to-Sequence Speech Recognition Systems. Proc. Interspeech 2019, 2220-2224, DOI: 10.21437/Interspeech.2019-2355.


@inproceedings{Karafiát2019,
  author={Martin Karafiát and Murali Karthick Baskar and Shinji Watanabe and Takaaki Hori and Matthew Wiesner and Jan Černocký},
  title={{Analysis of Multilingual Sequence-to-Sequence Speech Recognition Systems}},
  year=2019,
  booktitle={Proc. Interspeech 2019},
  pages={2220--2224},
  doi={10.21437/Interspeech.2019-2355},
  url={http://dx.doi.org/10.21437/Interspeech.2019-2355}
}