Back to Search Start Over

Example-Based Framework for Perceptually Guided Audio Texture Generation

Authors :
Kamath, Purnima
Gupta, Chitralekha
Wyse, Lonce
Nanayakkara, Suranga
Publication Year :
2023

Abstract

Controllable generation using StyleGANs is usually achieved by training the model using labeled data. For audio textures, however, there is currently a lack of large semantically labeled datasets. Therefore, to control generation, we develop a method for semantic control over an unconditionally trained StyleGAN in the absence of such labeled datasets. In this paper, we propose an example-based framework to determine guidance vectors for audio texture generation based on user-defined semantic attributes. Our approach leverages the semantically disentangled latent space of an unconditionally trained StyleGAN. By using a few synthetic examples to indicate the presence or absence of a semantic attribute, we infer the guidance vectors in the latent space of the StyleGAN to control that attribute during generation. Our results show that our framework can find user-defined and perceptually relevant guidance vectors for controllable generation for audio textures. Furthermore, we demonstrate an application of our framework to other tasks, such as selective semantic attribute transfer.<br />Comment: Accepted for publication at IEEE Transactions on Audio, Speech and Language Processing

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2308.11859
Document Type :
Working Paper
Full Text :
https://doi.org/10.1109/TASLP.2024.3393741