Utterance Verification for Text-Dependent Speaker Recognition: A Comparative Assessment Using the RedDots Corpus

Tomi Kinnunen, Md. Sahidullah, Ivan Kukanov, Héctor Delgado, Massimiliano Todisco, Achintya Kr. Sarkar, Nicolai Bæk Thomsen, Ville Hautamäki, Nicholas Evans, Zheng-Hua Tan


Text-dependent automatic speaker verification naturally calls for the simultaneous verification of speaker identity and spoken content. These two tasks can be achieved with automatic speaker verification (ASV) and utterance verification (UV) technologies. While both have been addressed previously in the literature, a treatment of simultaneous speaker and utterance verification with a modern, standard database is so far lacking. This is despite the burgeoning demand for voice biometrics in a plethora of practical security applications. With the goal of improving overall verification performance, this paper reports different strategies for simultaneous ASV and UV in the context of short-duration, text-dependent speaker verification. Experiments performed on the recently released RedDots corpus are reported for three different ASV systems and four different UV systems. Results show that the combination of utterance verification with automatic speaker verification is (almost) universally beneficial with significant performance improvements being observed.


DOI: 10.21437/Interspeech.2016-1125

Cite as

Kinnunen, T., Sahidullah, M., Kukanov, I., Delgado, H., Todisco, M., Sarkar, A.K., Thomsen, N.B., Hautamäki, V., Evans, N., Tan, Z. (2016) Utterance Verification for Text-Dependent Speaker Recognition: A Comparative Assessment Using the RedDots Corpus. Proc. Interspeech 2016, 430-434.

Bibtex
@inproceedings{Kinnunen+2016,
author={Tomi Kinnunen and Md. Sahidullah and Ivan Kukanov and Héctor Delgado and Massimiliano Todisco and Achintya Kr. Sarkar and Nicolai Bæk Thomsen and Ville Hautamäki and Nicholas Evans and Zheng-Hua Tan},
title={Utterance Verification for Text-Dependent Speaker Recognition: A Comparative Assessment Using the RedDots Corpus},
year=2016,
booktitle={Interspeech 2016},
doi={10.21437/Interspeech.2016-1125},
url={http://dx.doi.org/10.21437/Interspeech.2016-1125},
pages={430--434}
}