ISCA Archive Interspeech 2009
ISCA Archive Interspeech 2009

Trimmed KL divergence between Gaussian mixtures for robust unsupervised acoustic anomaly detection

Nash Borges, Gerard G. L. Meyer

In previous work [1], we presented several implementations of acoustic anomaly detection by training a model on purely normal data and estimating the divergence between it and other input. Here, we reformulate the problem in an unsupervised framework and allow for anomalous contamination of the training data. We focus exclusively on methods employing Gaussian mixture models (GMMs) since they are often used in speech processing systems. After analyzing what caused the Kullback-Leibler (KL) divergence between GMMs to break down in the face of training contamination, we came up with a promising solution. By trimming one quarter of the most divergent Gaussians from the mixture model, we significantly outperformed the untrimmed approximation for contamination levels of 10% and above, reducing the equal error rate from 33.8% to 6.4% at 33% contamination. The performance of the trimmed KL divergence showed no significant dependence on the investigated contamination levels.

N. Borges and G. G. L. Meyer, “Unsupervised distributional anomaly detection for a self-diagnostic speech activity detector,” in CISS, 2008.


doi: 10.21437/Interspeech.2009-673

Cite as: Borges, N., Meyer, G.G.L. (2009) Trimmed KL divergence between Gaussian mixtures for robust unsupervised acoustic anomaly detection. Proc. Interspeech 2009, 2555-2558, doi: 10.21437/Interspeech.2009-673

@inproceedings{borges09_interspeech,
  author={Nash Borges and Gerard G. L. Meyer},
  title={{Trimmed KL divergence between Gaussian mixtures for robust unsupervised acoustic anomaly detection}},
  year=2009,
  booktitle={Proc. Interspeech 2009},
  pages={2555--2558},
  doi={10.21437/Interspeech.2009-673}
}