An Investigation on Training Deep Neural Networks Using Probabilistic Transcriptions

Amit Das, Mark Hasegawa-Johnson


In this study, a transfer learning technique is presented for cross-lingual speech recognition in an adverse scenario where there are no natively transcribed transcriptions in the target language. The transcriptions that are available during training are transcribed by crowd workers who neither speak nor have any familiarity with the target language. Hence, such transcriptions are likely to be inaccurate. Training a deep neural network (DNN) in such a scenario is challenging; previously reported results have described DNN error rates exceeding the error rate of an adapted Gaussian Mixture Model (GMM). This paper investigates multi-task learning techniques using deep neural networks which are suitable for this scenario. We report, for the first time, absolute improvement in phone error rates (PER) in the range 1.3–6.2% over GMMs adapted to probabilistic transcriptions. Results are reported for Swahili, Hungarian, and Mandarin.


DOI: 10.21437/Interspeech.2016-655

Cite as

Das, A., Hasegawa-Johnson, M. (2016) An Investigation on Training Deep Neural Networks Using Probabilistic Transcriptions. Proc. Interspeech 2016, 3858-3862.

Bibtex
@inproceedings{Das+2016,
author={Amit Das and Mark Hasegawa-Johnson},
title={An Investigation on Training Deep Neural Networks Using Probabilistic Transcriptions},
year=2016,
booktitle={Interspeech 2016},
doi={10.21437/Interspeech.2016-655},
url={http://dx.doi.org/10.21437/Interspeech.2016-655},
pages={3858--3862}
}