The use of articulatory movement data in speech synthesis applications: An overview — Application of articulatory movements using machine learning algorithms

Korin Richmond, Zhenhua Ling, Junichi Yamagishi

Research output: Contribution to journalArticlepeer-review

Abstract / Description of output

This paper describes speech processing work in which articulator movements are used in conjunction with the acoustic speech signal and/or linguistic information. By ``articulator movements,'' we mean the changing positions of human speech articulators such as the tongue and lips, which may be recorded by electromagnetic articulography (EMA), amongst other articulography techniques. Specifically, we provide an overview of: i) inversion mapping techniques, where we estimate articulator movements from a given new speech waveform automatically; ii) statistical voice conversion and speech synthesis techniques which use articulator movements as part of the process to generate synthetic speech, and also make it intuitively controllable via articulation; and iii) automatic prediction (or synthesis) of articulator movements from any given new text input.
Original languageEnglish
Pages (from-to)467-477
Number of pages11
JournalAcoustical Science and Technology
Volume36
Issue number6
DOIs
Publication statusPublished - 1 Nov 2015

Fingerprint

Dive into the research topics of 'The use of articulatory movement data in speech synthesis applications: An overview — Application of articulatory movements using machine learning algorithms'. Together they form a unique fingerprint.

Cite this