The Speech Mapping concept posits that speech sequences can be represented by trajectories in a multiparametric space, whose elements are related to each other by relationships representing speech production mechanisms. This concept is viewed as a useful framework for pronunciation training, in a scheme where the teacher uses a Virtual Talking Head, to manipulate audio-visual speech stimuli in order to fulfil a double task: (1) evaluating and improving the learner's perception of the target language sounds, and (2) helping the learner produce the corresponding articulations by acquiring the internalisation of the relations between articulatory gestures and resulting sounds. We describe a set of data and models, including a virtual talking head, that can be useful for pronunciation training, present a few experiments supporting this approach, and suggest some directions for the future.
Cite as: Badin, P., Bailly, G., Boe, L.-J. (1998) Towards the use of a virtual talking head and of speech mapping tools for pronunciation training. Proc. ETRW on Speech Technology in Language Learning (STiLL), 167-170
@inproceedings{badin98_still, author={Pierre Badin and Gérard Bailly and Louis-Jean Boe}, title={{Towards the use of a virtual talking head and of speech mapping tools for pronunciation training}}, year=1998, booktitle={Proc. ETRW on Speech Technology in Language Learning (STiLL)}, pages={167--170} }