Visual Speech Synthesis Using Dynamic Visemes, Contextual Features and DNNs

Ausdang Thangthai, Ben Milner, Sarah Taylor


This paper examines methods to improve visual speech synthesis from a text input using a deep neural network (DNN). Two representations of the input text are considered, namely into phoneme sequences or dynamic viseme sequences. From these sequences, contextual features are extracted that include information at varying linguistic levels, from frame level down to the utterance level. These are extracted from a broad sliding window that captures context and produces features that are input into the DNN to estimate visual features. Experiments first compare the accuracy of these visual features against an HMM baseline method which establishes that both the phoneme and dynamic viseme systems perform better with best performance obtained by a combined phoneme-dynamic viseme system. An investigation into the features then reveals the importance of the frame level information which is able to avoid discontinuities in the visual feature sequence and produces a smooth and realistic output.


DOI: 10.21437/Interspeech.2016-1084

Cite as

Thangthai, A., Milner, B., Taylor, S. (2016) Visual Speech Synthesis Using Dynamic Visemes, Contextual Features and DNNs. Proc. Interspeech 2016, 2458-2462.

Bibtex
@inproceedings{Thangthai+2016,
author={Ausdang Thangthai and Ben Milner and Sarah Taylor},
title={Visual Speech Synthesis Using Dynamic Visemes, Contextual Features and DNNs},
year=2016,
booktitle={Interspeech 2016},
doi={10.21437/Interspeech.2016-1084},
url={http://dx.doi.org/10.21437/Interspeech.2016-1084},
pages={2458--2462}
}