Deep Neural Network Bottleneck Features for Acoustic Event Recognition

Seongkyu Mun, Suwon Shon, Wooil Kim, Hanseok Ko


Bottleneck features have been shown to be effective in improving the accuracy of speaker recognition, language identification and automatic speech recognition. However, few works have focused on bottleneck features for acoustic event recognition. This paper proposes a novel acoustic event recognition framework using bottleneck features derived from a Deep Neural Network (DNN). In addition to conventional features (MFCC, Mel-spectrum, etc.), this paper employs rhythm, timbre, and spectrum-statistics features for effectively extracting acoustic characteristics from audio signals. The effectiveness of the proposed method is demonstrated on a database of real life recordings via experiments, and its robust performance is verified by comparing to conventional methods.


DOI: 10.21437/Interspeech.2016-1112

Cite as

Mun, S., Shon, S., Kim, W., Ko, H. (2016) Deep Neural Network Bottleneck Features for Acoustic Event Recognition. Proc. Interspeech 2016, 2954-2957.

Bibtex
@inproceedings{Mun+2016,
author={Seongkyu Mun and Suwon Shon and Wooil Kim and Hanseok Ko},
title={Deep Neural Network Bottleneck Features for Acoustic Event Recognition},
year=2016,
booktitle={Interspeech 2016},
doi={10.21437/Interspeech.2016-1112},
url={http://dx.doi.org/10.21437/Interspeech.2016-1112},
pages={2954--2957}
}