Back to Search
Start Over
Example-Based Framework for Perceptually Guided Audio Texture Generation
- Publication Year :
- 2023
-
Abstract
- Controllable generation using StyleGANs is usually achieved by training the model using labeled data. For audio textures, however, there is currently a lack of large semantically labeled datasets. Therefore, to control generation, we develop a method for semantic control over an unconditionally trained StyleGAN in the absence of such labeled datasets. In this paper, we propose an example-based framework to determine guidance vectors for audio texture generation based on user-defined semantic attributes. Our approach leverages the semantically disentangled latent space of an unconditionally trained StyleGAN. By using a few synthetic examples to indicate the presence or absence of a semantic attribute, we infer the guidance vectors in the latent space of the StyleGAN to control that attribute during generation. Our results show that our framework can find user-defined and perceptually relevant guidance vectors for controllable generation for audio textures. Furthermore, we demonstrate an application of our framework to other tasks, such as selective semantic attribute transfer.<br />Comment: Accepted for publication at IEEE Transactions on Audio, Speech and Language Processing
Details
- Database :
- arXiv
- Publication Type :
- Report
- Accession number :
- edsarx.2308.11859
- Document Type :
- Working Paper
- Full Text :
- https://doi.org/10.1109/TASLP.2024.3393741