INTERSPEECH 2004 - ICSLP
8th International Conference on Spoken Language Processing
Jeju Island, Korea
October 4-8, 2004
From X-ray or MRU Data to Sounds through Articulatory Synthesis: towards an Integrated View of the Speech Communication Process
Laboratoire de Phonétique et de Phonologie,
Université Paris III, Paris, France
This tutorial presents an integrated method to simulate the transfer from X-ray (or MRI) data to
acoustics and finally to sounds. It illustrates the necessity of an articulatory model (hereby Maeda’s
model) so as to:
Construct realistic stimuli (sounds that human beings could really produce) for psychoacoustic
- "hear" what kind of sounds the vocal tract of a man or a woman, of a new-born or a monkey
could produce and inversely, what vocal shapes could produce a sound with given acoustic
- Study the correlation between the observed subtle articulatory and acoustic differences and the
choices of preferred prototypes in the realisation and perception of the same API symbol by
native speakers of different languages.
- Modelise vowels and consonants in context, and differentiate between transitional gestures
which are necessary in a co-articulation process, but not essential in order to differentiate
- Simulate the acoustic and perceptual consequences of the articulatory deformation realized by
the singers (e.g. singing formant), or in case of pathological voices.
Emphasis is put on the work done in our laboratory, and more generally by different teams in
France (Grenoble, Aix-en-Provence, Strasbourg and Nancy).
Vaissière, Jacqueline (2004):
"From x-ray or MRU data to sounds through articulatory synthesis: towards an integrated view of the speech communication process",
In INTERSPEECH-2004, paper P4.