VCV Synthesis Using Task Dynamics to Animate a Factor-Based Articulatory Model

Rachel Alexander, Tanner Sorensen, Asterios Toutios, Shrikanth S. Narayanan


This paper presents an initial architecture for articulatory synthesis which combines a dynamical system for the control of vocal tract shaping with a novel MATLAB implementation of an articulatory synthesizer. The dynamical system controls a speaker-specific vocal tract model derived by factor analysis of mid-sagittal real-time MRI data and provides input to the articulatory synthesizer, which simulates the propagation of sound waves in the vocal tract. First, parameters of the dynamical system are estimated from real-time MRI data of human speech production. Second, vocal-tract dynamics is simulated for vowel-consonant-vowel utterances using a sequence of two dynamical systems: the first one starts from a vowel vocal-tract configuration and achieves a vocal-tract closure; the second one starts from the closure and achieves the target configuration of the second vowel. Third, vocal-tract dynamics is converted to area function dynamics and is input to the synthesizer to generate the acoustic signal. Synthesized vowel-consonant-vowel examples demonstrate the feasibility of the method.


 DOI: 10.21437/Interspeech.2017-1410

Cite as: Alexander, R., Sorensen, T., Toutios, A., Narayanan, S.S. (2017) VCV Synthesis Using Task Dynamics to Animate a Factor-Based Articulatory Model. Proc. Interspeech 2017, 244-248, DOI: 10.21437/Interspeech.2017-1410.


@inproceedings{Alexander2017,
  author={Rachel Alexander and Tanner Sorensen and Asterios Toutios and Shrikanth S. Narayanan},
  title={VCV Synthesis Using Task Dynamics to Animate a Factor-Based Articulatory Model},
  year=2017,
  booktitle={Proc. Interspeech 2017},
  pages={244--248},
  doi={10.21437/Interspeech.2017-1410},
  url={http://dx.doi.org/10.21437/Interspeech.2017-1410}
}