Artigo Acesso aberto

Avocodo: Generative Adversarial Network for Artifact-Free Vocoder

2023; Association for the Advancement of Artificial Intelligence; Volume: 37; Issue: 11 Linguagem: Inglês

10.1609/aaai.v37i11.26479

ISSN

2374-3468

Autores

Taejun Bak, Junmo Lee, Hanbin Bae, Jinhyeok Yang, Jae‐Sung Bae, Young-Sun Joo,

Tópico(s)

Music and Audio Processing

Resumo

Neural vocoders based on the generative adversarial neural network (GAN) have been widely used due to their fast inference speed and lightweight networks while generating high-quality speech waveforms. Since the perceptually important speech components are primarily concentrated in the low-frequency bands, most GAN-based vocoders perform multi-scale analysis that evaluates downsampled speech waveforms. This multi-scale analysis helps the generator improve speech intelligibility. However, in preliminary experiments, we discovered that the multi-scale analysis which focuses on the low-frequency bands causes unintended artifacts, e.g., aliasing and imaging artifacts, which degrade the synthesized speech waveform quality. Therefore, in this paper, we investigate the relationship between these artifacts and GAN-based vocoders and propose a GAN-based vocoder, called Avocodo, that allows the synthesis of high-fidelity speech with reduced artifacts. We introduce two kinds of discriminators to evaluate speech waveforms in various perspectives: a collaborative multi-band discriminator and a sub-band discriminator. We also utilize a pseudo quadrature mirror filter bank to obtain downsampled multi-band speech waveforms while avoiding aliasing. According to experimental results, Avocodo outperforms baseline GAN-based vocoders, both objectively and subjectively, while reproducing speech with fewer artifacts.

Referência(s)