Robust Example Search Using Bottleneck Features for Example-Based Speech Enhancement

Atsunori Ogawa, Shogo Seki, Keisuke Kinoshita, Marc Delcroix, Takuya Yoshioka, Tomohiro Nakatani, Kazuya Takeda


Example-based speech enhancement is a promising approach for coping with highly non-stationary noise. Given a noisy speech input, it first searches in noisy speech corpora for the noisy speech examples that best match the input. Then, it concatenates the clean speech examples that are paired with the matched noisy examples to obtain an estimate of the underlying clean speech component in the input. This framework works well if the noisy speech corpora contain the noise included in the input. However, it is impossible to prepare corpora that cover all types of noisy environments. Moreover, the example search is usually performed using noise sensitive mel-frequency cepstral coefficient features (MFCCs). Consequently, a mismatch between an input and the corpora is inevitable. This paper proposes using bottleneck features (BNFs) extracted from a deep neural network (DNN) acoustic model for the example search. Since BNFs have good noise robustness (invariance), the mismatch is mitigated and thus a more accurate example search can be performed. Experimental results on the Aurora4 corpus show that the example-based approach using BNFs greatly improves the enhanced speech quality compared with that using MFCCs. It also consistently outperforms a conventional DNN-based approach, i.e. a denoising autoencoder.


DOI: 10.21437/Interspeech.2016-671

Cite as

Ogawa, A., Seki, S., Kinoshita, K., Delcroix, M., Yoshioka, T., Nakatani, T., Takeda, K. (2016) Robust Example Search Using Bottleneck Features for Example-Based Speech Enhancement. Proc. Interspeech 2016, 3733-3737.

Bibtex
@inproceedings{Ogawa+2016,
author={Atsunori Ogawa and Shogo Seki and Keisuke Kinoshita and Marc Delcroix and Takuya Yoshioka and Tomohiro Nakatani and Kazuya Takeda},
title={Robust Example Search Using Bottleneck Features for Example-Based Speech Enhancement},
year=2016,
booktitle={Interspeech 2016},
doi={10.21437/Interspeech.2016-671},
url={http://dx.doi.org/10.21437/Interspeech.2016-671},
pages={3733--3737}
}