LightGrad: Lightweight Diffusion Probabilistic Model for Text-to-speech

Accepted by ICASSP2023. Code is available at https://github.com/thuhcsi/LightGrad/ .

Abstract

Recent advances in neural text-to-speech (TTS) models bring thousands of TTS applications into daily life, where models are deployed in cloud to provide services for customs. Among these models are diffusion probabilistic models (DPMs), which can be stably trained and are more parameter-efficient compared with other generative models. As transmitting data between customs and the cloud introduce high latency and the risk of exposing private data, deploying TTS models on edge devices is preferred. When implementing DPMs onto edge devices, there are two practical problems. First, current DPMs are not lightweight enough for resource-constrained devices. Second, DPMs require many denoising steps in inference, which increases latency. In this work, we present LightGrad, a lightweight DPM for TTS. LightGrad is equipped with a lightweight U-Net diffusion decoder and a training-free fast sampling technique, reducing both model parameters and inference latency. Streaming inference is also implemented in LightGrad to reduce latency further. Compared with Grad-TTS, LightGrad achieves 62.2% reduction in paramters, 65.7% reduction in latency, while preserving comparable speech quality on both Chinese Mandarin and English in 4 denoising steps.

Experiments

We evaluated LightGrad on both Chinese and English dataset. For Chinese, we use a public speech dataset containing 10,000 audio clips whose total length is nearly 12 hours. Numbers of samples for training, validation and testing are 9,600, 200 and 200, respectively. For English, we use LJSpeech containing 13,100 English audio clips, whose total length is nearly 24 hours. Numbers of samples for training, validation and testing are 12,229, 348 and 523, respectively. Audios from both two datasets are resampled to 22,050Hz, and are converted to 80-dimensional mel-spectrograms with the frame size 1,024 and the hop size 256.

We select Grad-TTS as our baseline and follow its original setup. Our LightGrad consists of an encoder, a duration predictor and a lightweight U-Net decoder. The architecture of the encoder and the duration predictor is the same as Grad-TTS, but the encoder’s hidden size and number of convolution channels in encoder are 128 and 512 respectively. LightGrad is trained for 1.7M iterations on a single GPU with batch size 16, and Adam is chosen as the optimizer with learning rate 0.0001. T for the forward process of LightGrad is set to 1, and we use the same noise schedule as Grad-TTS in LightGrad. During inference, the temperature hyperparameter is set to 1.5 for both Grad-TTS and LightGrad. We use the number of function evaluations (NFE, a.k.a number of denoising steps) to represent the number of calls to the decoder when sampling from DPM, and we set NFE for Grad-TTS and LightGrad to 4 and 10. When LightGrad performs streaming inference, the decoder generates 0.5 second mel-spectrogram chunk each time. HiFi-GAN is chosen as the vocoder converting mel-spectrograms to audios.

To evaluate the speech quality of LightGrad, we conducted a subjective test to compare LightGrad with other systems, including speeches reconstructed from ground truth mel-spectrogram (GT(reconstructed)) and Grad-TTS. Mean opinion score (MOS) is selected as the evaluation metric of synthesized speeches. To obtain MOS, for each model we randomly select 20 samples from test set and present them to 20 subjects in random order. Subjects were asked to rate the quality of synthesized speech on a scale from 1 to 5 with 1 point increment in terms of naturalness, robustness and noise. Audios that are more natural, have fewer pronunciation mistakes and less noise, are considered better. We also conducted an objective evaluation using mel cepstral distortion (MCD). Additionally, a runtime performance comparison between LightGrad and Grad-TTS is performed. Average MCD, average latency, realtime factor (RTF) and runtime peak memory are calculated on the whole test set.

Experiment results are shown in Table.1.

Experiment Results

Demos from Chinese Dataset

In the following table, 10 synthesized sample pairs from the public Chinese speech dataset are provided.

Text
你看梁朝伟演色戒,刘嘉玲也得咬牙忍住呀.
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
面对这股勇猛的士气,师长牛七伟在感动之余格外镇静.
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
要不我们把预设方案上的敌情改一改?
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
这些聋哑人只有乖乖就范.
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
四百块钱,不花这四百,你永远挂不上号.
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
朱好龙摸了摸水牛,水牛乖乖睡了下来.
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
红圈处是设在鱼塘边的赌场.
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
我的闺蜜们一口咬定我是犯了花痴,看中了主人的帅气模样.
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
庭审时,两被告人在捂嘴咳嗽.
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
但阿灵顿现在却声称,她当初离职时处于极其肮脏的状况.
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming

Demos from English Dataset

In the following table, 10 synthesized sample pairs from LJSpeech are provided.

Text
after his death in the fourteen eighties , or at least by fourteen ninety , printing in venice had declined very much ;
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
yet their type is artistically on a much lower level than jensons , and in fact
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
and these somewhat wiry letters are suitable for the machine process , which would not do justice to letters of more generous design .
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
in reading the modern figures the eyes must be strained before the reader can have any reasonable assurance
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
for where these are boldly and carefully designed , and each letter is thoroughly individual in form ,
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
the words may be set much closer together , without loss of clearness .
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
the paper on which the printing is to be done is a necessary part of our subject :
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
the master felons side consisted of a yard the same size as the preceding , appropriated nominally to the most decent and better behaved prisoners ,
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
and the rest being provided with barrack beds , and in dimensions varying from thirty feet by fifteen to fifteen feet by ten .
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming
Text
the state of morals was disgraceful . any woman obtained admission if sober , and if she got drunk she was not turned out .
Ground Truth (reconstructed) Grad-TTS-10 LightGrad-10 Grad-TTS-4 LightGrad-4 LightGrad-4-Streaming

Ablation Study

To show the effectiveness of designs in LightGrad, ablation studies are conducted, where 20 subjects were asked to rate the comparison mean opinion score (CMOS) for 20 samples from test set in terms of naturalness, robustness and noise on both Chinese and English dataset. To validate the effectiveness of lightweight U-Net, we replace it with a different decoder composed of four feed-forward Transformer (FFT) block having roughly the same number of parameters as lightweight U-Net. We also substitute the fast sampling technique in LightGrad with the original sampling method in Grad-TTS to show benefits of the fast sampling technique. NFE for all diffusion models in ablation studies is set to 4.

Demos from on Chinese Dataset

In the following table, 6 synthesized sample pairs from the public Chinese speech dataset are provided.

Text
你看梁朝伟演色戒,刘嘉玲也得咬牙忍住呀.
LightGrad-4 w/o diffusion w/o fast sampling
Text
这些聋哑人只有乖乖就范.
LightGrad-4 w/o diffusion w/o fast sampling
Text
四百块钱,不花这四百,你永远挂不上号.
LightGrad-4 w/o diffusion w/o fast sampling
Text
我的闺蜜们一口咬定我是犯了花痴,看中了主人的帅气模样.
LightGrad-4 w/o diffusion w/o fast sampling
Text
但阿灵顿现在却声称,她当初离职时处于极其肮脏的状况.
LightGrad-4 w/o diffusion w/o fast sampling
Text
是全党全国各族人民团结奋斗的旗帜?
LightGrad-4 w/o diffusion w/o fast sampling

Demos from English Dataset

In the following table, 6 synthesized sample pairs from LJSpeech dataset are provided.

Text
that the forms of printed letters should follow more or less closely those of the written character , and they followed them very closely .
LightGrad-4 w/o diffusion w/o fast sampling
Text
after his death in the fourteen eighties , or at least by fourteen ninety , printing in venice had declined very much ;
LightGrad-4 w/o diffusion w/o fast sampling
Text
yet their type is artistically on a much lower level than jensons , and in fact
LightGrad-4 w/o diffusion w/o fast sampling
Text
and are worthy representatives of the eager enthusiasm for the revived learning of that epoch . by far ,
LightGrad-4 w/o diffusion w/o fast sampling
Text
and what remained was divided into a masters and a common side for female felons .
LightGrad-4 w/o diffusion w/o fast sampling
Text
with jurisdiction to hold pleas in all actions within the prescribed limits . the court was chiefly used for the recovery of small debts under ten pounds
LightGrad-4 w/o diffusion w/o fast sampling