Deep Auto-Encoder Based Multi-Task Learning Using Probabilistic Transcriptions

Amit Das, Mark Hasegawa-Johnson, Karel Veselý


We examine a scenario where we have no access to native transcribers in the target language. This is typical of language communities that are under-resourced. However, turkers (online crowd workers) available in online marketplaces can serve as valuable alternative resources for providing transcripts in the target language. We assume that the turkers neither speak nor have any familiarity with the target language. Thus, they are unable to distinguish all phone pairs in the target language; their transcripts therefore specify, at best, a probability distribution called a probabilistic transcript (PT). Standard deep neural network (DNN) training using PTs do not necessarily improve error rates. Previously reported results have demonstrated some success by adopting the multi-task learning (MTL) approach. In this study, we report further improvements by introducing a deep auto-encoder based MTL. This method leverages large amounts of untranscribed data in the target language in addition to the PTs obtained from turkers. Furthermore, to encourage transfer learning in the feature space, we also examine the effect of using monophones from transcripts in well-resourced languages. We report consistent improvement in phone error rates (PER) for Swahili, Amharic, Dinka, and Mandarin.


 DOI: 10.21437/Interspeech.2017-582

Cite as: Das, A., Hasegawa-Johnson, M., Veselý, K. (2017) Deep Auto-Encoder Based Multi-Task Learning Using Probabilistic Transcriptions. Proc. Interspeech 2017, 2073-2077, DOI: 10.21437/Interspeech.2017-582.


@inproceedings{Das2017,
  author={Amit Das and Mark Hasegawa-Johnson and Karel Veselý},
  title={Deep Auto-Encoder Based Multi-Task Learning Using Probabilistic Transcriptions},
  year=2017,
  booktitle={Proc. Interspeech 2017},
  pages={2073--2077},
  doi={10.21437/Interspeech.2017-582},
  url={http://dx.doi.org/10.21437/Interspeech.2017-582}
}