Back to Search Start Over

Investigating the Viability of Masked Language Modeling for Symbolic Music Generation in abc-notation

Authors :
Casini, Luca
Jonason, Nicolas
Sturm, Bob
Casini, Luca
Jonason, Nicolas
Sturm, Bob
Publication Year :
2024

Abstract

The dominating approach for modeling sequences (e.g. text, music) with deep learning is the causal approach, which consists in learning to predict tokens sequentially given those preceding it. Another paradigm is masked language modeling, which consists of learning to predict the masked tokens of a sequence in no specific order, given all non-masked tokens. Both approaches can be used for generation, but the latter is more flexible for editing, e.g. changing the middle of a sequence. This paper investigates the viability of masked language modeling applied to Irish traditional music represented in the text-based format abc-notation. Our model, called abcMLM, enables a user to edit tunes in arbitrary ways while retaining similar generation capabilities to causal models. We find that generation using masked language modeling is more challenging, but leveraging additional information from a dataset, e.g., imputing musical structure, can generate sequences that are on par with previous models.<br />QC 20240604Part of ISBN 978-3-031-56991-3; 978-3-031-56992-0

Details

Database :
OAIster
Notes :
English
Publication Type :
Electronic Resource
Accession number :
edsoai.on1457577440
Document Type :
Electronic Resource
Full Text :
https://doi.org/10.1007.978-3-031-56992-0_6