Neural Attention Models for Sequence Classification: Analysis and Application to Key Term Extraction and Dialogue Act Detection

Sheng-syun Shen, Hung-Yi Lee


Recurrent neural network architectures combining with attention mechanism, or neural attention model, have shown promising performance recently for the tasks including speech recognition, image caption generation, visual question answering and machine translation. In this paper, neural attention model is applied on two sequence labeling tasks, dialogue act detection and key term extraction. In the sequence labeling tasks, the model input is a sequence, and the output is the label of the input sequence. The major difficulty of sequence labeling is that when the input sequence is long, it can include many noisy or irrelevant part. If the information in the whole sequence is treated equally, the noisy or irrelevant part may degrade the classification performance. The attention mechanism is helpful for sequence classification task because it is capable of highlighting important part among the entire sequence for the classification task. The experimental results show that with the attention mechanism, discernible improvements were achieved in the sequence labeling task considered here. The roles of the attention mechanism in the tasks are further analyzed and visualized in this paper.


DOI: 10.21437/Interspeech.2016-1359

Cite as

Shen, S., Lee, H. (2016) Neural Attention Models for Sequence Classification: Analysis and Application to Key Term Extraction and Dialogue Act Detection. Proc. Interspeech 2016, 2716-2720.

Bibtex
@inproceedings{Shen+2016,
author={Sheng-syun Shen and Hung-Yi Lee},
title={Neural Attention Models for Sequence Classification: Analysis and Application to Key Term Extraction and Dialogue Act Detection},
year=2016,
booktitle={Interspeech 2016},
doi={10.21437/Interspeech.2016-1359},
url={http://dx.doi.org/10.21437/Interspeech.2016-1359},
pages={2716--2720}
}