Annotator Trustability-based Cooperative Learning Solutions for Intelligent Audio Analysis

Simone Hantke, Christoph Stemp, Björn Schuller


A broad range of artificially intelligent applications are nowadays available resulting in a need for masses of labelled data for the underlying machine learning models. This annotated data, however, is scarce and expensive to obtain from expert-like annotators. Crowdsourcing has been shown as a viable alternative, but it has to be carried out with adequate quality control to obtain reliable labels. Whilst crowdsourcing allows for the rapid collection of large-scale annotations, another technique called Cooperative Learning, aims at reducing the overall annotation costs, by learning to select only the most important instances for manual annotation. In this regard, we investigate the advantages of this approach and combine crowdsourcing with different iterative cooperative learning paradigms for audio data annotation, incorporating an annotator trustability score to reduce the labelling effort needed and, at the same time, to achieve better classification results. Key experimental results on an emotion recognition task show a considerable relative annotation reduction compared to a ‘non-intelligent’ approach of up to 85.3%. Moreover, the proposed trustability-based methods reach an unweighted average recall of 74.8%, while the baseline approach peaks at 61.2%. Therefore, the proposed trustability-based approaches efficiently reduce the manual annotation load, as well as improving the model.


 DOI: 10.21437/Interspeech.2018-1019

Cite as: Hantke, S., Stemp, C., Schuller, B. (2018) Annotator Trustability-based Cooperative Learning Solutions for Intelligent Audio Analysis. Proc. Interspeech 2018, 3504-3508, DOI: 10.21437/Interspeech.2018-1019.


@inproceedings{Hantke2018,
  author={Simone Hantke and Christoph Stemp and Björn Schuller},
  title={Annotator Trustability-based Cooperative Learning Solutions for Intelligent Audio Analysis},
  year=2018,
  booktitle={Proc. Interspeech 2018},
  pages={3504--3508},
  doi={10.21437/Interspeech.2018-1019},
  url={http://dx.doi.org/10.21437/Interspeech.2018-1019}
}