Conversational Analysis Using Utterance-level Attention-based Bidirectional Recurrent Neural Networks

Chandrakant Bothe, Sven Magg, Cornelius Weber, Stefan Wermter


Recent approaches for dialogue act recognition have shown that context from preceding utterances is important to classify the subsequent one. It was shown that the performance improves rapidly when the context is taken into account. We propose an utterance-level attention-based bidirectional recurrent neural network (Utt-Att-BiRNN) model to analyze the importance of preceding utterances to classify the current one. In our setup, the BiRNN is given the input set of current and preceding utterances. Our model outperforms previous models that use only preceding utterances as context on the used corpus. Another contribution of our research is a mechanism to discover the amount of information in each utterance to classify the subsequent one and to show that context-based learning not only improves the performance but also achieves higher confidence in the recognition of dialogue acts. We use character- and word-level features to represent the utterances. The results are presented for character and word feature representations and as an ensemble model of both representations. We found that when classifying short utterances, the closest preceding utterances contribute to a higher degree.


 DOI: 10.21437/Interspeech.2018-2527

Cite as: Bothe, C., Magg, S., Weber, C., Wermter, S. (2018) Conversational Analysis Using Utterance-level Attention-based Bidirectional Recurrent Neural Networks. Proc. Interspeech 2018, 996-1000, DOI: 10.21437/Interspeech.2018-2527.


@inproceedings{Bothe2018,
  author={Chandrakant Bothe and Sven Magg and Cornelius Weber and Stefan Wermter},
  title={Conversational Analysis Using Utterance-level Attention-based Bidirectional Recurrent Neural Networks},
  year=2018,
  booktitle={Proc. Interspeech 2018},
  pages={996--1000},
  doi={10.21437/Interspeech.2018-2527},
  url={http://dx.doi.org/10.21437/Interspeech.2018-2527}
}