Preprint Article Version 1 This version is not peer-reviewed

Advancements in End-to-End Audio Style Transformation: A Differentiable Approach for Voice Conversion and Musical Style Transfer

Version 1 : Received: 5 November 2024 / Approved: 6 November 2024 / Online: 6 November 2024 (12:58:36 CET)

How to cite: Aggarwal, S.; Uttam, S.; Garg, S.; Garg, S.; Jain, K.; Aggarwal, S. Advancements in End-to-End Audio Style Transformation: A Differentiable Approach for Voice Conversion and Musical Style Transfer. Preprints 2024, 2024110396. https://doi.org/10.20944/preprints202411.0396.v1 Aggarwal, S.; Uttam, S.; Garg, S.; Garg, S.; Jain, K.; Aggarwal, S. Advancements in End-to-End Audio Style Transformation: A Differentiable Approach for Voice Conversion and Musical Style Transfer. Preprints 2024, 2024110396. https://doi.org/10.20944/preprints202411.0396.v1

Abstract

We introduce a fully differentiable end-to-end audio transformation network designed to convert the style of one audio sample to another. This method offers three significant advantages: (a) it operates without the need for parallel utterances, transcriptions, or time alignment processes; (b) it utilizes a global conditioning mechanism, making it vocabulary agnostic and capable of transforming audio styles regardless of the target identity; and (c) it performs one-shot audio transformations without intermediate phonetic representations, thus eliminating the necessity for phonetic alignments and speaker-independent ASR networks. We assess our method against existing approaches in voice conversion and musical style transfer tasks. Subjective evaluations demonstrate the superiority of our approach. The network employs an encoder-decoder architecture that integrates neural network models known for their explainability in natural language processing tasks.

Keywords

Voice Conversion; Musical Style Transfer; Audio Transformations; End-to-End Audio Pipeline

Subject

Computer Science and Mathematics, Artificial Intelligence and Machine Learning

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.