Learning Interpretable Control Dimensions for Speech Synthesis by Using External Data

Zack Hodari, Oliver Watts, Srikanth Ronanki, Simon King


There are many aspects of speech that we might want to control when creating text-to-speech (TTS) systems. We present a general method that enables control of arbitrary aspects of speech, which we demonstrate on the task of emotion control. Current TTS systems use supervised machine learning and are therefore heavily reliant on labelled data. If no labels are available for a desired control dimension, then creating interpretable control becomes challenging. We introduce a method that uses external, labelled data (i.e. not the original data used to train the acoustic model) to enable the control of dimensions that are not labelled in the original data. Adding interpretable control allows the voice to be manually controlled to produce more engaging speech, for applications such as audiobooks. We evaluate our method using a listening test.


 DOI: 10.21437/Interspeech.2018-2075

Cite as: Hodari, Z., Watts, O., Ronanki, S., King, S. (2018) Learning Interpretable Control Dimensions for Speech Synthesis by Using External Data. Proc. Interspeech 2018, 32-36, DOI: 10.21437/Interspeech.2018-2075.


@inproceedings{Hodari2018,
  author={Zack Hodari and Oliver Watts and Srikanth Ronanki and Simon King},
  title={Learning Interpretable Control Dimensions for Speech Synthesis by Using External Data},
  year=2018,
  booktitle={Proc. Interspeech 2018},
  pages={32--36},
  doi={10.21437/Interspeech.2018-2075},
  url={http://dx.doi.org/10.21437/Interspeech.2018-2075}
}