Document-Level Neural TTS Using Curriculum Learning and Attention Masking
Speech synthesis has been developed to the level of natural human-level speech synthesized through an attention-based end-to-end text-to-speech synthesis (TTS) model. However, it is difficult to generate attention when synthesizing a text longer than the trained length or document-level text. In thi...
Saved in:
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2021-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9312676/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1832582411817320448 |
---|---|
author | Sung-Woong Hwang Joon-Hyuk Chang |
author_facet | Sung-Woong Hwang Joon-Hyuk Chang |
author_sort | Sung-Woong Hwang |
collection | DOAJ |
description | Speech synthesis has been developed to the level of natural human-level speech synthesized through an attention-based end-to-end text-to-speech synthesis (TTS) model. However, it is difficult to generate attention when synthesizing a text longer than the trained length or document-level text. In this paper, we propose a neural speech synthesis model that can synthesize more than 5 min of speech at once using training data comprising a short speech of less than 10 s. This model can be used for tasks that need to synthesize document-level speech at a time, such as a singing voice synthesis (SVS) system or a book reading system. First, through curriculum learning, our model automatically increases the length of the speech trained for each epoch, while reducing the batch size so that long sentences can be trained with a limited graphics processing unit (GPU) capacity. During synthesis, the document-level text is synthesized using only the necessary contexts of the current time step and masking the rest through an attention-masking mechanism. The Tacotron2-based speech synthesis model and duration predictor were used in the experiment, and the results showed that proposed method can synthesize document-level speech with overwhelmingly lower character error rate, and attention error rates, and higher quality than those obtained using the existing model. |
format | Article |
id | doaj-art-e5e9d71aa603485cab3e50756fab497d |
institution | Kabale University |
issn | 2169-3536 |
language | English |
publishDate | 2021-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Access |
spelling | doaj-art-e5e9d71aa603485cab3e50756fab497d2025-01-30T00:00:58ZengIEEEIEEE Access2169-35362021-01-0198954896010.1109/ACCESS.2020.30490739312676Document-Level Neural TTS Using Curriculum Learning and Attention MaskingSung-Woong Hwang0https://orcid.org/0000-0001-6194-9752Joon-Hyuk Chang1https://orcid.org/0000-0003-2610-2323Department of Electronic Engineering, Hanyang University, Seoul, South KoreaDepartment of Electronic Engineering, Hanyang University, Seoul, South KoreaSpeech synthesis has been developed to the level of natural human-level speech synthesized through an attention-based end-to-end text-to-speech synthesis (TTS) model. However, it is difficult to generate attention when synthesizing a text longer than the trained length or document-level text. In this paper, we propose a neural speech synthesis model that can synthesize more than 5 min of speech at once using training data comprising a short speech of less than 10 s. This model can be used for tasks that need to synthesize document-level speech at a time, such as a singing voice synthesis (SVS) system or a book reading system. First, through curriculum learning, our model automatically increases the length of the speech trained for each epoch, while reducing the batch size so that long sentences can be trained with a limited graphics processing unit (GPU) capacity. During synthesis, the document-level text is synthesized using only the necessary contexts of the current time step and masking the rest through an attention-masking mechanism. The Tacotron2-based speech synthesis model and duration predictor were used in the experiment, and the results showed that proposed method can synthesize document-level speech with overwhelmingly lower character error rate, and attention error rates, and higher quality than those obtained using the existing model.https://ieeexplore.ieee.org/document/9312676/Speech synthesisdocument-level neural TTScurriculum learningattention maskingTacotron2MelGAN |
spellingShingle | Sung-Woong Hwang Joon-Hyuk Chang Document-Level Neural TTS Using Curriculum Learning and Attention Masking IEEE Access Speech synthesis document-level neural TTS curriculum learning attention masking Tacotron2 MelGAN |
title | Document-Level Neural TTS Using Curriculum Learning and Attention Masking |
title_full | Document-Level Neural TTS Using Curriculum Learning and Attention Masking |
title_fullStr | Document-Level Neural TTS Using Curriculum Learning and Attention Masking |
title_full_unstemmed | Document-Level Neural TTS Using Curriculum Learning and Attention Masking |
title_short | Document-Level Neural TTS Using Curriculum Learning and Attention Masking |
title_sort | document level neural tts using curriculum learning and attention masking |
topic | Speech synthesis document-level neural TTS curriculum learning attention masking Tacotron2 MelGAN |
url | https://ieeexplore.ieee.org/document/9312676/ |
work_keys_str_mv | AT sungwoonghwang documentlevelneuralttsusingcurriculumlearningandattentionmasking AT joonhyukchang documentlevelneuralttsusingcurriculumlearningandattentionmasking |