Self-Attentional Acoustic Models

Matthias Sperber, Jan Niehues, Graham Neubig, Sebastian Stüker, Alex Waibel


Self-attention is a method of encoding sequences of vectors by relating these vectors to each-other based on pairwise similarities. These models have recently shown promising results for modeling discrete sequences, but they are non-trivial to apply to acoustic modeling due to computational and modeling issues. In this paper, we apply self-attention to acoustic modeling, proposing several improvements to mitigate these issues: First, self-attention memory grows quadratically in the sequence length, which we address through a downsampling technique. Second, we find that previous approaches to incorporate position information into the model are unsuitable and explore other representations and hybrid models to this end. Third, to stress the importance of local context in the acoustic signal, we propose a Gaussian biasing approach that allows explicit control over the context range. Experiments find that our model approaches a strong baseline based on LSTMs with network-in-network connections while being much faster to compute. Besides speed, we find that interpretability is a strength of self-attentional acoustic models and demonstrate that self-attention heads learn a linguistically plausible division of labor.


 DOI: 10.21437/Interspeech.2018-1910

Cite as: Sperber, M., Niehues, J., Neubig, G., Stüker, S., Waibel, A. (2018) Self-Attentional Acoustic Models. Proc. Interspeech 2018, 3723-3727, DOI: 10.21437/Interspeech.2018-1910.


@inproceedings{Sperber2018,
  author={Matthias Sperber and Jan Niehues and Graham Neubig and Sebastian Stüker and Alex Waibel},
  title={Self-Attentional Acoustic Models},
  year=2018,
  booktitle={Proc. Interspeech 2018},
  pages={3723--3727},
  doi={10.21437/Interspeech.2018-1910},
  url={http://dx.doi.org/10.21437/Interspeech.2018-1910}
}