Papers
arxiv:2506.04013

Towards Better Disentanglement in Non-Autoregressive Zero-Shot Expressive Voice Conversion

Published on Jun 4
Authors:
,
,

Abstract

The model enhances voice conversion by reducing timbre leakage and improving expressivity transfer through multilingual discrete speech units, augmentation-based similarity loss, mix-style layer normalization, and enriched style embeddings with global pitch and energy features.

AI-generated summary

Expressive voice conversion aims to transfer both speaker identity and expressive attributes from a target speech to a given source speech. In this work, we improve over a self-supervised, non-autoregressive framework with a conditional variational autoencoder, focusing on reducing source timbre leakage and improving linguistic-acoustic disentanglement for better style transfer. To minimize style leakage, we use multilingual discrete speech units for content representation and reinforce embeddings with augmentation-based similarity loss and mix-style layer normalization. To enhance expressivity transfer, we incorporate local F0 information via cross-attention and extract style embeddings enriched with global pitch and energy features. Experiments show our model outperforms baselines in emotion and speaker similarity, demonstrating superior style adaptation and reduced source style leakage.

Community

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2506.04013 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2506.04013 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2506.04013 in a Space README.md to link it from this page.

Collections including this paper 1