Factorised Representations for Neural Network Adaptation to Diverse Acoustic Environments

Joachim Fainberg, Steve Renals, Peter Bell


Adapting acoustic models jointly to both speaker and environment has been shown to be effective. In many realistic scenarios, however, either the speaker or environment at test time might be unknown, or there may be insufficient data to learn a joint transform. Generating independent speaker and environment transforms improves the match of an acoustic model to unseen combinations. Using i-vectors, we demonstrate that it is possible to factorise speaker or environment information using multi-condition training with neural networks. Specifically, we extract bottleneck features from networks trained to classify either speakers or environments. We perform experiments on the Wall Street Journal corpus combined with environment noise from the Diverse Environments Multichannel Acoustic Noise Database. Using the factorised i-vectors we show improvements in word error rates on perturbed versions of the eval92 and dev93 test sets, both when one factor is missing and when the factors are seen but not in the desired combination.


 DOI: 10.21437/Interspeech.2017-1365

Cite as: Fainberg, J., Renals, S., Bell, P. (2017) Factorised Representations for Neural Network Adaptation to Diverse Acoustic Environments. Proc. Interspeech 2017, 749-753, DOI: 10.21437/Interspeech.2017-1365.


@inproceedings{Fainberg2017,
  author={Joachim Fainberg and Steve Renals and Peter Bell},
  title={Factorised Representations for Neural Network Adaptation to Diverse Acoustic Environments},
  year=2017,
  booktitle={Proc. Interspeech 2017},
  pages={749--753},
  doi={10.21437/Interspeech.2017-1365},
  url={http://dx.doi.org/10.21437/Interspeech.2017-1365}
}