Edinburgh Research Explorer

Learning interpretable control dimensions for speech synthesis by using external data

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Original languageEnglish
Title of host publicationInterspeech 2018
Place of PublicationHyderabad, India
Number of pages5
Publication statusPublished - 6 Sep 2018
EventInterspeech 2018 - Hyderabad International Convention Centre, Hyderabad, India
Duration: 2 Sep 20186 Sep 2018


ConferenceInterspeech 2018
Internet address


There are many aspects of speech that we might want to control when creating text-to-speech (TTS) systems. We present a general method that enables control of arbitrary aspects of speech, which we
demonstrate on the task of emotion control. Current TTS systems use supervised machine learning and are therefore heavily reliant on labelled data. If no labels are available for a desired control dimension, then creating interpretable control becomes challenging. We introduce a method that uses external, labelled data (i.e. not the original data used to train the acoustic model) to enable the control of dimensions that are not labelled in the original data. Adding interpretable control allows the voice to be manually controlled to produce more engaging speech, for applications such as audiobooks. We evaluate our method using a
listening test.


Interspeech 2018


Hyderabad, India

Event: Conference

Download statistics

No data available

ID: 74303443