Real-Time Tracking of Speakers’ Emotions, States, and Traits on Mobile Platforms

Erik Marchi, Florian Eyben, Gerhard Hagerer, Björn Schuller


We demonstrate audEERING’s sensAI technology running natively on low-resource mobile devices applied to emotion analytics and speaker characterisation tasks. A show-case application for the Android platform is provided, where audEERING’s highly noise robust voice activity detection based on LSTM-RNN is combined with our core emotion recognition and speaker characterisation engine natively on the mobile device. This eliminates the need for network connectivity and allows to perform robust speaker state and trait recognition efficiently in real-time without network transmission lags. Real-time factors are benchmarked for a popular mobile device to demonstrate the efficiency, and average response times are compared to a server based approach. The output of the emotion analysis is visualized graphically in the arousal and valence space alongside the emotion category and further speaker characteristics.


Cite as

Marchi, E., Eyben, F., Hagerer, G., Schuller, B. (2016) Real-Time Tracking of Speakers’ Emotions, States, and Traits on Mobile Platforms. Proc. Interspeech 2016, 1182-1183.

Bibtex
@inproceedings{Marchi+2016,
author={Erik Marchi and Florian Eyben and Gerhard Hagerer and Björn Schuller},
title={Real-Time Tracking of Speakers’ Emotions, States, and Traits on Mobile Platforms},
year=2016,
booktitle={Interspeech 2016},
pages={1182--1183}
}