Unsupervised Adaptation with Adversarial Dropout Regularization for Robust Speech Recognition

Pengcheng Guo, Sining Sun, Lei Xie


Recent adversarial methods proposed for unsupervised domain adaptation of acoustic models try to fool a specific domain discriminator and learn both senone-discriminative and domain-invariant hidden feature representations. However, a drawback of these approaches is that the feature generator simply aligns different features into the same distribution without considering the class boundaries of the target domain data. Thus, ambiguous target domain features can be generated near the decision boundaries, decreasing speech recognition performance. In this study, we propose to use Adversarial Dropout Regularization (ADR) in acoustic modeling to overcome the foregoing issue. Specifically, we optimize the senone classifier to make its decision boundaries lie in the class boundaries of unlabeled target data. Then, the feature generator learns to create features far away from the decision boundaries, which are more discriminative. We apply the ADR approach on the CHiME-3 corpus and the proposed method yields up to 12.9% relative WER reductions compared with the baseline trained on source domain data only and further improvement over the widely used gradient reversal layer method.


 DOI: 10.21437/Interspeech.2019-2544

Cite as: Guo, P., Sun, S., Xie, L. (2019) Unsupervised Adaptation with Adversarial Dropout Regularization for Robust Speech Recognition. Proc. Interspeech 2019, 749-753, DOI: 10.21437/Interspeech.2019-2544.


@inproceedings{Guo2019,
  author={Pengcheng Guo and Sining Sun and Lei Xie},
  title={{Unsupervised Adaptation with Adversarial Dropout Regularization for Robust Speech Recognition}},
  year=2019,
  booktitle={Proc. Interspeech 2019},
  pages={749--753},
  doi={10.21437/Interspeech.2019-2544},
  url={http://dx.doi.org/10.21437/Interspeech.2019-2544}
}