Back to Search Start Over

ChapGTP, ILLC's Attempt at Raising a BabyLM: Improving Data Efficiency by Automatic Task Formation

Authors :
Jumelet, Jaap
Hanna, Michael
Kloots, Marianne de Heer
Langedijk, Anna
Pouw, Charlotte
van der Wal, Oskar
Publication Year :
2023

Abstract

We present the submission of the ILLC at the University of Amsterdam to the BabyLM challenge (Warstadt et al., 2023), in the strict-small track. Our final model, ChapGTP, is a masked language model that was trained for 200 epochs, aided by a novel data augmentation technique called Automatic Task Formation. We discuss in detail the performance of this model on the three evaluation suites: BLiMP, (Super)GLUE, and MSGS. Furthermore, we present a wide range of methods that were ultimately not included in the model, but may serve as inspiration for training LMs in low-resource settings.<br />Comment: Part of the BabyLM challenge at CoNLL

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2310.11282
Document Type :
Working Paper