The Voice Conversion Challenge 2018: Promoting Development of Parallel and Nonparallel Methods

Jaime Lorenzo-Trueba, Junichi Yamagishi, Tomoki Toda, Daisuke Saito, Fernando Villavicencio, Tomi Kinnunen, Zhenhua Ling

Research output: Chapter in Book/Report/Conference proceedingConference contribution


We present the Voice Conversion Challenge 2018, designed as a follow up to the 2016 edition with the aim of providing a common framework for evaluating and comparing different state-of-the-art voice conversion (VC) systems. The objective of the challenge was to perform speaker conversion (i.e. transform the vocal identity) of a source speaker to a target speaker while maintaining linguistic information. As an update to the previous challenge, we considered both parallel and nonparallel data to form the Hub and Spoke tasks, respectively. A total of 23 teams from around the world submitted their systems, 11 of them additionally participated in the optional Spoke task. A large-scale crowdsourced perceptual evaluation was then carried out to rate the submitted converted speech in terms of naturalness and similarity to the target speaker identity. In this paper, we present a brief summary of the state-of-the-art techniques for VC, followed by a detailed explanation of the challenge tasks and the results that were obtained.
Original languageEnglish
Title of host publicationSpeaker Odyssey 2018
Subtitle of host publicationThe Speaker and Language Recognition Workshop
Place of Publication Les Sables d’Olonne, France
Number of pages8
Publication statusPublished - 29 Jun 2018
EventThe Speaker and Language Recognition Workshop - Les Sables d’Olonne, France
Duration: 26 Jun 201829 Jun 2018

Publication series

ISSN (Electronic)2312-2846


ConferenceThe Speaker and Language Recognition Workshop
Abbreviated titleSpeaker Odyssey 2018
CityLes Sables d’Olonne
Internet address


Dive into the research topics of 'The Voice Conversion Challenge 2018: Promoting Development of Parallel and Nonparallel Methods'. Together they form a unique fingerprint.

Cite this