Interspeech'2005 - Eurospeech

Lisbon, Portugal
September 4-8, 2005

Articulatory Motivated Acoustic Features for Speech Recognition

Daniil Kocharov (1), András Zolnay (2), Ralf Schlüter (2), Hermann Ney (2)

(1) Saint-Petersburg State University, Russia; (2) RWTH Aachen University, Germany

In this paper, we consider the use of multiple acoustic features of the speech signal for continuous speech recognition. A novel articulatory motivated acoustic feature is introduced, namely the spectrum derivative feature. The new feature is tested in combination with the standard Mel Frequency Cepstral Coefficients (MFCC) and the voicedness features. Linear Discriminant Analysis is applied to find the optimal combination of different acoustic features. Experiments have been performed on small and large vocabulary tasks. Significant improvements in word error rate have been obtained by combining the MFCC feature with the articulatory motivated voicedness and spectrum derivative features: improvements of up to 25% on the small-vocabulary task and improvements of up to 4% on the large-vocabulary task relative to using MFCC alone with the same overall number of parameters in the system.

Full Paper

Bibliographic reference.  Kocharov, Daniil / Zolnay, András / Schlüter, Ralf / Ney, Hermann (2005): "Articulatory motivated acoustic features for speech recognition", In INTERSPEECH-2005, 1101-1104.