Lombard speech synthesis using transfer learning in a Tacotron text-to-speech system

Loading...
Thumbnail Image
Access rights
openAccess
Journal Title
Journal ISSN
Volume Title
Conference article in proceedings
This publication is imported from Aalto University research portal.
View publication in the Research portal
View/Open full text file from the Research portal
Date
2019
Major/Subject
Mcode
Degree programme
Language
en
Pages
2833-2837
Series
Proceedings of Interspeech, Interspeech - Annual Conference of the International Speech Communication Association
Abstract
Currently, there is increasing interest to use sequence-to-sequence models in text-to-speech (TTS) synthesis with attention like that in Tacotron models. These models are end-to-end, meaning that they learn both co-articulation and duration properties directly from text and speech. Since these models are entirely data-driven, they need large amounts of data to generate synthetic speech of good quality. However, in challenging speaking styles, such as Lombard speech, it is difficult to record sufficiently large speech corpora. Therefore, we propose a transfer learning method to adapt a TTS system of normal speaking style to Lombard style. We also experiment with a WaveNet vocoder along with a traditional vocoder (WORLD) in the synthesis of Lombard speech. The subjective and objective evaluation results indicated that the proposed adaptation system coupled with the WaveNet vocoder clearly outperformed the conventional deep neural network based TTS system in the synthesis of Lombard speech
Description
Keywords
Adaptation, Lombard speaking style, Tacotron, Text-To-Speech (TTS)
Other note
Citation
Bollepalli , B , Juvela , L & Alku , P 2019 , Lombard speech synthesis using transfer learning in a Tacotron text-to-speech system . in Proceedings of Interspeech . Interspeech - Annual Conference of the International Speech Communication Association , International Speech Communication Association (ISCA) , pp. 2833-2837 , Interspeech , Graz , Austria , 15/09/2019 . https://doi.org/10.21437/Interspeech.2019-1333