Discriminating Languages in a Probabilistic Latent Subspace

Aleksandr Sizov, Kong Aik Lee, Tomi Kinnunen


We explore a method to boost discriminative capabilities of Probabilistic Linear Discriminant Analysis (PLDA) model without losing its generative advantages. To this end, our focus is in a low-dimensional PLDA latent subspace. We optimize the model with respect to MMI (Maximum Mutual Information) and our own objective functions, which is an approximation to the detection cost function. We evaluate the performance on NIST Language Recognition Evaluation 2015. Our model trains faster and performs more accurately in comparison to both generative PLDA and discriminative LDA baselines with 12% and 4% relative improvement in the average detection cost, respectively. The proposed method is applicable for a broad range of closed-set tasks.


DOI: 10.21437/Odyssey.2016-12

Cite as

Sizov, A., Lee, K.A., Kinnunen, T. (2016) Discriminating Languages in a Probabilistic Latent Subspace. Proc. Odyssey 2016, 81-88.

Bibtex
@inproceedings{Sizov+2016,
author={Aleksandr Sizov and Kong Aik Lee and Tomi Kinnunen},
title={Discriminating Languages in a Probabilistic Latent Subspace},
year=2016,
booktitle={Odyssey 2016},
doi={10.21437/Odyssey.2016-12},
url={http://dx.doi.org/10.21437/Odyssey.2016-12},
pages={81--88}
}