INTERSPEECH 2008
9th Annual Conference of the International Speech Communication Association

Brisbane, Australia
September 22-26, 2008

Pronunciation Training: The Role of Eye and Ear

Dominic W. Massaro (1), Stephanie Bigler (1), Trevor Chen (1), Marcus Perlman (1), Slim Ouni (2)

(1) University of California at Santa Cruz, USA; (2) LORIA, France

For speech perception and production of a new language, we examined whether 1) they would be more easily learned by ear and eye relative to by ear alone, and 2) whether viewing the tongue, palate, and velum during production is more beneficial for learning than a standard frontal view of the speaker. In addition, we determine whether differences in learning under these conditions are due to enhanced receptive learning from additional visual information, or to more active learning motivated by the visual presentations. Test stimuli were two similar vowels in Mandarin and two similar stop consonants in Arabic, presented in different word contexts. Participants were tested with auditory speech and were either trained 1) unimodally with just auditory speech or bimodally with both auditory and visual speech, and 2) a standard frontal view versus an inside view of the vocal tract. The visual speech was generated by the appropriate multilingual versions of Baldi [1]. The results test the effectiveness of visible speech for learning a new language. Preliminary results indicate that visible speech can contribute positively to acquiring new speech distinctions and promoting active learning.

Reference

  1. Massaro, D. W. (1998). Perceiving talking faces: From speech perception to a behavioral principle. Cambridge, Massachusetts: MIT Press.

Full Paper

Bibliographic reference.  Massaro, Dominic W. / Bigler, Stephanie / Chen, Trevor / Perlman, Marcus / Ouni, Slim (2008): "Pronunciation training: the role of eye and ear", In INTERSPEECH-2008, 2623-2626.