ISCA Archive Interspeech 2009
ISCA Archive Interspeech 2009

Impact of different speaking modes on EMG-based speech recognition

Michael Wand, Szu-Chen Stan Jou, Arthur R. Toth, Tanja Schultz

We present our recent results on speech recognition by surface electromyography (EMG), which captures the electric potentials that are generated by the human articulatory muscles. This technique can be used to enable Silent Speech Interfaces, since EMG signals are generated even when people only articulate speech without producing any sound. Preliminary experiments have shown that the EMG signals created by audible and silent speech are quite distinct. In this paper we first compare various methods of initializing a silent speech EMG recognizer, showing that the performance of the recognizer substantially varies across different speakers. Based on this, we analyze EMG signals from audible and silent speech, present first results on how discrepancies between these speaking modes affect EMG recognizers, and suggest areas for future work.


doi: 10.21437/Interspeech.2009-228

Cite as: Wand, M., Jou, S.-C.S., Toth, A.R., Schultz, T. (2009) Impact of different speaking modes on EMG-based speech recognition. Proc. Interspeech 2009, 648-651, doi: 10.21437/Interspeech.2009-228

@inproceedings{wand09_interspeech,
  author={Michael Wand and Szu-Chen Stan Jou and Arthur R. Toth and Tanja Schultz},
  title={{Impact of different speaking modes on EMG-based speech recognition}},
  year=2009,
  booktitle={Proc. Interspeech 2009},
  pages={648--651},
  doi={10.21437/Interspeech.2009-228}
}