ISCA Archive SLTU 2014
ISCA Archive SLTU 2014

Towards real-time multilingual multimodal speech-to-speech translation

Satoshi Nakamura

Speech-to-speech translation technology enables natural oral communication between different language speaking people. Many research projects have addressed speech-to-speech translation (S2ST) technology, such as ATR, VERBMOBIL, C-STAR, NESPOLE!, BABYLON, GALE, and EU-bridge. The speechto- speech translation system is normally composed of automatic speech recognition (ASR), machine translation (MT), and speech synthesis (TTS). All of the modules are corpus-based and statistical model-based systems. In this talk, new challenges toward a real-time multimodal speechto- speech translation will be introduced.

Index Terms: Speech-to-speech translation, S2ST, multimodal processing, multilingual systems


Cite as: Nakamura, S. (2014) Towards real-time multilingual multimodal speech-to-speech translation. Proc. 4th Workshop on Spoken Language Technologies for Under-Resourced Languages (SLTU 2014), 13-15

@inproceedings{nakamura14_sltu,
  author={Satoshi Nakamura},
  title={{Towards real-time multilingual multimodal speech-to-speech translation}},
  year=2014,
  booktitle={Proc. 4th Workshop on Spoken Language Technologies for Under-Resourced Languages  (SLTU 2014)},
  pages={13--15}
}