ISCA Archive

Auditory-Visual Speech Processing (AVSP 2001)

September 7-9, 2001
Aalborg, Denmark

Edited by Dominic W. Massaro,
Joanna Light, and Kristin Geraci



Bibliographic Reference

[AVSP-2001] Auditory-Visual Speech Processing (AVSP 2001), ed. by Dominic W. Massaro, Joanna Light, and Kristin Geraci, Aalborg, Denmark, September 7-9, 2001, ISCA Archive, http://www.isca_speech.org/archive_open/avsp01



Author Index and Quick Access to Abstracts

Abry   Alwan   Andersen (167)   Andersen (172)   Anderson   Arb   Auer (7)   Auer (40)   Auer (50)   Auer (104)   Badin   Bailly   Bangham   Baum   Bernstein (7)   Bernstein (40)   Bernstein (50)   Bernstein (104)   Berthommier (183)   Berthommier (189)   Brammer   Bohning   Bothe   Burnham   Callan, A.   Callan, D.   Calvert   Campbell (13)   Campbell (36)   Campbell (199b)   Campbell (199c)   Cathiard   Cawley   Choi   Clark   Cohen   Colin   Cosi   Cuddy   Davis   Deltenre   Dodd   Ekvall   Elisei   Ellis   Faulkner   Goecke   Gowdy   Grant   Greenberg   Gurbuz   Gustafson   Hardison   Hazan   Heckmann   Jiang   Jordan   Kabisch   Karmiloff-Smith   Kim, Jeesun   Kim, Jinyoung   Kroos   Kroschel   Kshirsagar   Kuratate   Lampinen   Lau   Lee   Lidestam   Lin   Lyberg   Lyxell   MacSweeney   Magnenat-Thalmann   Magno-Caldognetto   Masuda   Massaro   Matthews   Mattys   McCotter   Millar   Morishima   Munhall   Nakamura   Neti   Nicholson   Nisch   Odisio   Ogata   Ortega-Llebaria   Ouhyoung   Patterson   Pelachaud   Ponton   Potamianos   Radeau   Randén   Robert-Ribes   Rothkrantz   Sams (167)   Sams (172)   Savariaux   Schoknecht   Schwartz (18)   Schwartz (138)   Singh   Slyh   Stiefelhagen   Straube   Tam   Theobald   Tiippana (167)   Tiippana (172)   Tufekci   Vatikiotis-Bateson (24)   Vatikiotis-Bateson (45)   Waibel   Wojdel   Yang   Yeh   Zelinsky   Zmarich  

Names written in boldface refer to first authors. Please note that each abstract opens in a separate window.



Table of Contents and Access to Abstracts

Visible Speech for Animation and Speechreading by Humans

Lidestam, Björn / Lyxell, Björn: "Speechreading essentials: signal, paralinguistic cues, and skill", 1-6.

Auer Jr., Edward T. / Bernstein, Lynne E. / Mattys, Sven: "The INFLUENCE OF THE LEXICON ON VISUAL SPOKEN WORD RECognition", 7-12.

Ellis, Tara / MacSweeney, Mairead / Dodd, Barbara / Campbell, Ruth: "TAS: A new test of adult speechreading - deaf people really can be better speechreaders", 13-17.

Schwartz, Jean-Luc / Savariaux, Christophe: "Is it easier to lipread one's own speech gestures than those of somebody else? it seems not!", 18-23.

Kroos, Christian / Masuda, Saeko / Kuratate, Takaaki / Vatikiotis-Bateson, Eric: "Towards the facecoder: dynamic face synthesis based on image motion estimation in speech", 24-29.

Kshirsagar, Sumedha / Magnenat-Thalmann, Nadia: "Viseme space for realistic speech animation", 30-35.

Brain Activation in Auditory Visual Processing

Bohning, M. / Campbell, Ruth / Karmiloff-Smith, A.: "Audiovisual speech perception in Williams Syndrome ", 36-39.

Auer Jr., Edward T. / Bernstein, Lynne E. / Singh, Manbir: "Comparing cortical activity during the perception of two forms of biological motion for language communication", 40-44.

Callan, Daniel / Callan, Akiko / Vatikiotis-Bateson, Eric: "Neural areas underlying the processing of visual speech information under conditions of degraded auditory information", 45-49.

Bernstein, Lynne E. / Jiang, Jintao / Alwan, Abeer / Auer Jr., Edward T.: "Similarity structure in visual phonetic perception and optical phonetics", 50-55.

Colin, C. / Radeau, M. / Deltenre, P.: "The mismatch negativity (MMN) and the McGurk effect", 56-61.

Nicholson, Karen / Baum, Shari / Cuddy, Lola / Munhall, Kevin: "A case of multimodal aprosodia: impaired auditory and visual speech prosody perception in a patient with right hemisphere damage", 62-65.

Facial Animation and Visual Speech Synthesis

Lin, I-Chen / Yeh, Jeng-Sheng / Ouhyoung, Ming: "Extraction of 3D facial motion parameters from mirror-reflected multi-view video for audio-visual synthesis", 66-71.

Pelachaud, C. / Magno-Caldognetto, E. / Zmarich, C. / Cosi, P.: "Modelling an Italian talking head ", 72-77.

Theobald, Barry J. / Bangham, J. Andrew / Matthews, Iain / Cawley, Gavin C.: "Visual speech synthesis using statistical models of shape and appearance", 78-83.

Arb, Allan / Gustafson, Steven / Anderson, Timothy / Slyh, Raymond: "Hidden Markov models for visual speech synthesis with limited data", 84-89.

Elisei, F. / Odisio, M. / Bailly, Gérard / Badin, Pierre: "Creating and controlling video-realistic talking heads", 90-97.

Morishima, Shigeo / Ogata, Shin / Nakamura, Satoshi: "Multimodal translation", 98-103.

Correlates of Auditory and Visual Speech

Bernstein, Lynne E. / Ponton, Curtis W. / Auer Jr., Edward T.: "Electrophysiology of unimodal and audiovisual speech perception", 104-109.

Kim, Jinyoung / Choi, Seungho / Lee, Joohun: "Development of a lip-sync algorithm based on an audio-visual corpus", 110-114.

Goecke, Roland / Millar, J. Bruce / Zelinsky, Alexander / Robert-Ribes, Jordi: "Analysis of audio-video correlation in vowels in Australian English", 115-120.

Ekvall, Christel / Lyberg, Bertil / Randén, Michael: "Non-verbal correlates to focal accents in Swedish", 121-126.

Kim, Jeesun / Davis, Chris: "Visible speech cues and auditory detection of spoken sentences: an effect of degree of correlation between acoustic and visual properties ", 127-131.

Grant, Ken W. / Greenberg, Steven: "Speech intelligibility derived from asynchronous processing of auditory-visual information", 132-137.

Auditory Visual Speech Perception by Humans

Cathiard, M.A. / Schwartz, Jean-Luc / Abry, C.: "Asking a naive question about the McGurk effect: Why does audio [b] give more [d] percepts with visual [g] than with visual [d]?", 138-142.

McCotter, M.V. / Jordan, T.R.: "Investigating the role of luminance boundaries in visual and audiovisual speech recognition using line drawn faces", 143-148.

Ortega-Llebaria, M. / Faulkner, A. / Hazan, Valerie: "Auditory-visual L2 speech perception: Effects of visual cues and acoustic-phonetic context for Spanish learners of English", 149-154.

Burnham, Denis / Lau, Susanna / Tam, Helen / Schoknecht, Colin: "Visual discrimination of cantonese tone by tonal but non-Cantonese speakers, and by non-tonal language speakers ", 155-160.

Hardison, Debra M.: "Bimodal word identification: effects of modality, speech style, sentence and phonetic/visual context", 161-166.

Tiippana, K. / Sams, M. / Andersen, T. S.: "Visual attention influences audiovisual speech perception", 167-171.

Auditory Visual Speech Recognition by Humans and by Machine

Andersen, T.S. / Tiippana, K. / Lampinen, J. / Sams, M.: "Modeling of audiovisual speech perception in noise", 172-176.

Potamianos, Gerasimos / Neti, Chalapathy: "Automatic speechreading of impaired speech", 177-182.

Berthommier, Frederic: "Audio-visual recognition of spectrally reduced speech", 183-188.

Heckmann, Martin / Berthommier, Frederic / Kroschel, Kristian: "A hybrid ANN/HMM audio-visual speech recognition system", 189-194.

Patterson, E. K. / Gurbuz, S. / Tufekci, Z. / Gowdy, J. N.: "Noise-based audio-visual fusion for robust speech recognition", 195-198.

Poster Presentations (no full papers available)

Bothe, Hans-Heinrich: "LIPPS - A visual telephone for hearing-impaired", 199.

Calvert, G. A. / Brammer, M. J. / Campbell, Ruth: "Cortical substrates of seeing speech: still and moving faces", 199.

Kabisch, Bjorn / Nisch, Carol / Straube, Eckart R. / Campbell, Ruth: "Development of a completely computerized McGurk design under variation of the signal to noise ratio", 199.

Stiefelhagen, Rainer / Yang, Jie / Waibel, Alex: "Estimating focus of attention based on gaze and sound", 200.

Wojdel, Jacek C. / Rothkrantz, Leon J.M.: "Obtaining person-independent feature space for lip reading", 200.

Cohen, Michael M. / Clark, Rashid / Massaro, Dominic W.: "Animated speech: research progress and applications", 200.