INTERSPEECH 2006 - ICSLP
The aim of this work is to examine the correlation between audio and visual speech features. The motivation is to find visual features that can provide clean audio feature estimates which can be used for speech enhancement when the original audio signal is corrupted by noise. Two audio features (MFCCs and formants) and three visual features (active appearance model, 2-D DCT and cross-DCT) are considered with correlation measured using multiple linear regression. The correlation is then exploited through the development of a maximum a posteriori (MAP) prediction of audio features solely from the visual features. Experiments reveal that features representing broad spectral information have higher correlation to visual features than those representing finer spectral detail. The accuracy of prediction follows the results found in the correlation measurements.
Bibliographic reference. Almajai, Ibrahim / Milner, Ben / Darch, Jonathan (2006): "Analysis of correlation between audio and visual speech features for clean audio feature prediction in noise", In INTERSPEECH-2006, paper 1634-Thu2WeO.5.