Tracking Contours of Orofacial Articulators from Real-Time MRI of Speech

Mathieu Labrunie, Pierre Badin, Dirk Voit, Arun A. Joseph, Laurent Lamalle, Coriandre Vilain, Louis-Jean Boë, Jens Frahm

We introduce a method for predicting midsagittal contours of orofacial articulators from real-time MRI data. A corpus of about 26 minutes of speech has been recorded of a French speaker at a rate of 55 images / s using highly undersampled radial gradient-echo MRI with image reconstruction by nonlinear inversion. The contours of each articulator have been manually traced for a set of about 60 images selected — by hierarchical clustering — to optimally represent the diversity of the speaker articulations. The data serve to build articulator-specific Principal Component Analysis (PCA) models of contours and associated image intensities, as well as multilinear regression (MLR) models that predict contour parameters from image parameters. The contours obtained by MLR are then refined, using the local information about pixel intensity profiles along the contours’ normals, by means of modified Active Shape Models (ASM) trained on the same data. The method reaches RMS of predicted points to reference contour distances between 0.54 and 0.93 mm, depending on articulators. The processing of the corpus demonstrated the efficiency of the procedure, despite the possibility of further improvements. This work opens new perspectives for studying articulatory motion in speech.

DOI: 10.21437/Interspeech.2016-78

Cite as

Labrunie, M., Badin, P., Voit, D., Joseph, A.A., Lamalle, L., Vilain, C., Boë, L., Frahm, J. (2016) Tracking Contours of Orofacial Articulators from Real-Time MRI of Speech. Proc. Interspeech 2016, 470-474.

author={Mathieu Labrunie and Pierre Badin and Dirk Voit and Arun A. Joseph and Laurent Lamalle and Coriandre Vilain and Louis-Jean Boë and Jens Frahm},
title={Tracking Contours of Orofacial Articulators from Real-Time MRI of Speech},
booktitle={Interspeech 2016},