8th International Conference on Spoken Language Processing

Jeju Island, Korea
October 4-8, 2004

From X-ray or MRU Data to Sounds through Articulatory Synthesis: towards an Integrated View of the Speech Communication Process

Jacqueline Vaissière

Laboratoire de Phonétique et de Phonologie, Université Paris III, Paris, France

This tutorial presents an integrated method to simulate the transfer from X-ray (or MRI) data to acoustics and finally to sounds. It illustrates the necessity of an articulatory model (hereby Maeda’s model) so as to:

    1. Construct realistic stimuli (sounds that human beings could really produce) for psychoacoustic experiments.
    2. "hear" what kind of sounds the vocal tract of a man or a woman, of a new-born or a monkey could produce and inversely, what vocal shapes could produce a sound with given acoustic characteristics.
    3. Study the correlation between the observed subtle articulatory and acoustic differences and the choices of preferred prototypes in the realisation and perception of the same API symbol by native speakers of different languages.
    4. Modelise vowels and consonants in context, and differentiate between transitional gestures which are necessary in a co-articulation process, but not essential in order to differentiate phonemes.
    5. Simulate the acoustic and perceptual consequences of the articulatory deformation realized by the singers (e.g. singing formant), or in case of pathological voices.
      1. Emphasis is put on the work done in our laboratory, and more generally by different teams in France (Grenoble, Aix-en-Provence, Strasbourg and Nancy).

        Full Paper

        Bibliographic reference.  Vaissière, Jacqueline (2004): "From x-ray or MRU data to sounds through articulatory synthesis: towards an integrated view of the speech communication process", In INTERSPEECH-2004, paper P4.