RetrieverTTS: Modeling Decomposed Factors for Text-Based Speech Insertion

Published in InterSpeech, 2022

[Paper] [demo] [code]

Architecture

This paper proposes a new “decompose-and-edit” paradigm for the text-based speech insertion task that facilitates arbitrary-length speech insertion and even full sentence generation. In the proposed paradigm, global and local factors in speech are explicitly decomposed and separately manipulated to achieve high speaker similarity and continuous prosody. Specifically, we proposed to represent the global factors by multiple tokens, which are extracted by cross-attention operation and then injected back by link-attention operation. Due to the rich representation of global factors, we manage to achieve high speaker similarity in a zero-shot manner. In addition, we introduce a prosody smoothing task to make the local prosody factor context-aware and therefore achieve satisfactory prosody continuity. We further achieve high voice quality with an adversarial training stage. In the subjective test, our method achieves state-of-the-art performance in both naturalness and similarity.

Cite

@article{yin2022retrievertts,
  title={RetrieverTTS: Modeling Decomposed Factors for Text-Based Speech Insertion},
  author={Dacheng, Yin and Chuanxin, Tang and Yanqing, Liu and Xiaoqiang, Wang and Zhiyuan, Zhao and Yucheng, Zhao and Zhiwei, Xiong and Sheng, Zhao and Chong, Luo}
  booktitle={Interspeech},
  year={2022}
}