A Hierarchical Attention Network-Based Approach for Depression Detection from Transcribed Clinical Interviews

Adria Mallol-Ragolta, Ziping Zhao, Lukas Stappen, Nicholas Cummins, Björn W. Schuller


The high prevalence of depression in society has given rise to a need for new digital tools that can aid its early detection. Among other effects, depression impacts the use of language. Seeking to exploit this, this work focuses on the detection of depressed and non-depressed individuals through the analysis of linguistic information extracted from transcripts of clinical interviews with a virtual agent. Specifically, we investigated the advantages of employing hierarchical attention-based networks for this task. Using Global Vectors (GloVe) pretrained word embedding models to extract low-level representations of the words, we compared hierarchical local-global attention networks and hierarchical contextual attention networks. We performed our experiments on the Distress Analysis Interview Corpus - Wizard of Oz (DAIC-WoZ) dataset, which contains audio, visual, and linguistic information acquired from participants during a clinical session. Our results using the DAIC-WoZ test set indicate that hierarchical contextual attention networks are the most suitable configuration to detect depression from transcripts. The configuration achieves an Unweighted Average Recall (UAR) of .66 using the test set, surpassing our baseline, a Recurrent Neural Network that does not use attention.


 DOI: 10.21437/Interspeech.2019-2036

Cite as: Mallol-Ragolta, A., Zhao, Z., Stappen, L., Cummins, N., Schuller, B.W. (2019) A Hierarchical Attention Network-Based Approach for Depression Detection from Transcribed Clinical Interviews. Proc. Interspeech 2019, 221-225, DOI: 10.21437/Interspeech.2019-2036.


@inproceedings{Mallol-Ragolta2019,
  author={Adria Mallol-Ragolta and Ziping Zhao and Lukas Stappen and Nicholas Cummins and Björn W. Schuller},
  title={{A Hierarchical Attention Network-Based Approach for Depression Detection from Transcribed Clinical Interviews}},
  year=2019,
  booktitle={Proc. Interspeech 2019},
  pages={221--225},
  doi={10.21437/Interspeech.2019-2036},
  url={http://dx.doi.org/10.21437/Interspeech.2019-2036}
}