Analysis and Optimization of Bottleneck Features for Speaker Recognition

Alicia Lozano-Diez, Anna Silnova, Pavel Matejka, Ondrej Glembek, Oldrich Plchot, Jan Pesan, Lukas Burget, Joaquin Gonzalez-Rodriguez

Recently, Deep Neural Network (DNN) based bottleneck features proved to be very effective in i-vector based speaker recognition. However, the bottleneck feature extraction is usually fully optimized for speech rather than speaker recognition task. In this paper, we explore whether DNNs suboptimal for speech recognition can provide better bottleneck features for speaker recognition. We experiment with different features optimized for speech or speaker recognition as input to the DNN. We also experiment with under-trained DNN, where the training was interrupted before the full convergence of the speech recognition objective. Moreover, we analyze the effect of normalizing the features at the input and/or at the output of bottleneck features extraction to see how it affects the final speaker recognition system performance. We evaluated the systems in the SRE’10, condition 5, female task. Results show that the best configuration of the DNN in terms of phone accuracy does not necessary imply better performance of the final speaker recognition system. Finally, we compare the performance of bottleneck features and the standard MFCC features in i-vector/PLDA speaker recognition system. The best bottleneck features yield up to 37% of relative improvement in terms of EER.

DOI: 10.21437/Odyssey.2016-51

Cite as

Lozano-Diez, A., Silnova, A., Matejka, P., Glembek, O., Plchot, O., Pesan, J., Burget, L., Gonzalez-Rodriguez, J. (2016) Analysis and Optimization of Bottleneck Features for Speaker Recognition. Proc. Odyssey 2016, 352-357.

author={Alicia Lozano-Diez and Anna Silnova and Pavel Matejka and Ondrej Glembek and Oldrich Plchot and Jan Pesan and Lukas Burget and Joaquin Gonzalez-Rodriguez},
title={Analysis and Optimization of Bottleneck Features for Speaker Recognition},
booktitle={Odyssey 2016},