Improved Example-Based Speech Enhancement by Using Deep Neural Network Acoustic Model for Noise Robust Example Search

Atsunori Ogawa, Keisuke Kinoshita, Marc Delcroix, Tomohiro Nakatani


Example-based speech enhancement is a promising single-channel approach for coping with highly nonstationary noise. Given a noisy speech input, it first searches in a noisy speech corpus for the noisy speech examples that best match the input. Then, it concatenates the clean speech examples that are paired with the matched noisy examples to obtain an estimate of the underlying clean speech component in the input. The quality of the enhanced speech depends on how accurate an example search can be performed given a noisy speech input. The example search is conventionally performed using a Gaussian mixture model (GMM) with mel-frequency cepstral coefficient features (MFCCs). To improve the noise robustness of the GMM-based example search, instead of using noise sensitive MFCCs, we have proposed using bottleneck features (BNFs), which are extracted from a deep neural network-based acoustic model (DNN-AM) built for automatic speech recognition. In this paper, instead of using a GMM with noise robust BNFs, we propose the direct use of a DNN-AM in the example search to further improve its noise robustness. Experimental results on the Aurora4 corpus show that the DNN-AM-based example search steadily improves the enhanced speech quality compared with the GMM-based example search using BNFs.


 DOI: 10.21437/Interspeech.2017-543

Cite as: Ogawa, A., Kinoshita, K., Delcroix, M., Nakatani, T. (2017) Improved Example-Based Speech Enhancement by Using Deep Neural Network Acoustic Model for Noise Robust Example Search. Proc. Interspeech 2017, 1963-1967, DOI: 10.21437/Interspeech.2017-543.


@inproceedings{Ogawa2017,
  author={Atsunori Ogawa and Keisuke Kinoshita and Marc Delcroix and Tomohiro Nakatani},
  title={Improved Example-Based Speech Enhancement by Using Deep Neural Network Acoustic Model for Noise Robust Example Search},
  year=2017,
  booktitle={Proc. Interspeech 2017},
  pages={1963--1967},
  doi={10.21437/Interspeech.2017-543},
  url={http://dx.doi.org/10.21437/Interspeech.2017-543}
}