Virtual Adversarial Training and Data Augmentation for Acoustic Event Detection with Gated Recurrent Neural Networks

Matthias Zöhrer, Franz Pernkopf


In this paper, we use gated recurrent neural networks (GRNNs) for efficiently detecting environmental events of the IEEE Detection and Classification of Acoustic Scenes and Events challenge (DCASE2016). For this acoustic event detection task data is limited. Therefore, we propose data augmentation such as on-the-fly shuffling and virtual adversarial training for regularization of the GRNNs. Both improve the performance using GRNNs. We obtain a segment-based error rate of 0.59 and an F-score of 58.6%.


 DOI: 10.21437/Interspeech.2017-1238

Cite as: Zöhrer, M., Pernkopf, F. (2017) Virtual Adversarial Training and Data Augmentation for Acoustic Event Detection with Gated Recurrent Neural Networks. Proc. Interspeech 2017, 493-497, DOI: 10.21437/Interspeech.2017-1238.


@inproceedings{Zöhrer2017,
  author={Matthias Zöhrer and Franz Pernkopf},
  title={Virtual Adversarial Training and Data Augmentation for Acoustic Event Detection with Gated Recurrent Neural Networks},
  year=2017,
  booktitle={Proc. Interspeech 2017},
  pages={493--497},
  doi={10.21437/Interspeech.2017-1238},
  url={http://dx.doi.org/10.21437/Interspeech.2017-1238}
}