INTERSPEECH 2009
10th Annual Conference of the International Speech Communication Association

Brighton, United Kingdom
September 6-10, 2009

Emotion Classification in Children's Speech Using Fusion of Acoustic and Linguistic Features

Tim Polzehl (1), Shiva Sundaram (1), Hamed Ketabdar (1), Michael Wagner (2), Florian Metze (3)

(1) Technische Universität Berlin, Germany
(2) University of Canberra, Australia
(3) Carnegie Mellon University, USA

This paper describes a system to detect angry vs. non-angry utterances of children who are engaged in dialog with an Aibo robot dog. The system was submitted to the Interspeech2009 Emotion Challenge evaluation. The speech data consist of short utterances of the children’s speech, and the proposed system is designed to detect anger in each given chunk. Frame-based cepstral features, prosodic and acoustic features as well as glottal excitation features are extracted automatically, reduced in dimensionality and classified by means of an artificial neural network and a support vector machine. An automatic speech recognizer transcribes the words in an utterance and yields a separate classification based on the degree of emotional salience of the words. Late fusion is applied to make a final decision on anger vs. non-anger of the utterance. Preliminary results show 75.9% unweighted average recall on the training data and 67.6% on the test set.

Full Paper

Bibliographic reference.  Polzehl, Tim / Sundaram, Shiva / Ketabdar, Hamed / Wagner, Michael / Metze, Florian (2009): "Emotion classification in children's speech using fusion of acoustic and linguistic features", In INTERSPEECH-2009, 340-343.