7th International Conference on Spoken Language Processing

September 16-20, 2002
Denver, Colorado, USA

Auditory-Visual Speech Perception Examined by Brain Imaging and Reaction Time

Kaoru Sekiyama (1), Yoichi Sugita (2)

(1) Future University Hakodate, Japan; (2) National Institute of Advanced Industrial Science and Technology, Japan

By using the McGurk effect [1], we compared brain activation during audiovisual (AV) speech perception for two sets of conditions differing in the intelligibility of auditory speech (High vs. Low). In the Low intelligibility condition in which speech was harder to hear, the McGurk effect, the visual influence, was much stronger. Functional magnetic resonance imaging (fMRI) also showed that speechreadingrelated visual areas (the left MT and left intraparietal sulcus as observed in the video-only condition) were strongly activated in the Low intelligibility AV condition but not in the High intelligibility AV condition. Thus visual information of the mouth movements was processed more intensively when speech was harder to hear. Reaction time data suggested that when auditory speech is easier to hear, there is a top-down suppression of visual processing that starts earlier than auditory processing. On the other hand, when auditory speech was less intelligible, reaction time data were such that visual mouth movements served as a priming cue. These results provide an insight into a time-spanned scope of the integration process.

Full Paper

Bibliographic reference.  Sekiyama, Kaoru / Sugita, Yoichi (2002): "Auditory-visual speech perception examined by brain imaging and reaction time", In ICSLP-2002, 1693-1696.