This report describes a three-year project, now eight months old, to develop interactive learning tools for language training with profoundly deaf children. The tools combine four key technologies: speech recognition, developed at the Oregon Graduate Institute; speech synthesis, developed at the University of Edinburgh and modified at OGI; facial animation, developed at University of California, Santa Cruz; and face tracking and speech reading, developed at Carnegie Mellon University. These technologies are being combined to create an intelligent conversational agent; a three-dimensional face that produces and understands auditory and visual speech. The agent has been incorporated into the CSLU Toolkit, a software environment for developing and researching spoken language systems. We describe our experiences in bringing interactive learning tools to classrooms at the Tucker-Maxon Oral School in Portland, Oregon, and the technological advances that are required for this project to succeed.
Cite as: Cole, R., Carmell, T., Connors, P., Macon, M., Wouters, J., Villiers, J.d., Tarachow, A., Massaro, D., Cohen, M., Beskow, J., Yang, J., Meier, U., Waibel, A., Stone, P., Fortier, G., Davis, A., Soland, C. (1998) Intelligent animated agents for interactive language training. Proc. ETRW on Speech Technology in Language Learning (STiLL), 163-166
@inproceedings{cole98_still, author={Ron Cole and Tim Carmell and Pom Connors and Mike Macon and Johan Wouters and Jacques de Villiers and Alice Tarachow and Dominic Massaro and Michael Cohen and Jonas Beskow and Jie Yang and Uwe Meier and Alex Waibel and Pat Stone and George Fortier and Alice Davis and Chris Soland}, title={{Intelligent animated agents for interactive language training}}, year=1998, booktitle={Proc. ETRW on Speech Technology in Language Learning (STiLL)}, pages={163--166} }