Back to Search Start Over

Distractor Generation Through Text-to-Text Transformer Models

Authors :
David De-Fitero-Dominguez
Eva Garcia-Lopez
Antonio Garcia-Cabot
Jesus-Angel Del-Hoyo-Gabaldon
Antonio Moreno-Cediel
Source :
IEEE Access, Vol 12, Pp 25580-25589 (2024)
Publication Year :
2024
Publisher :
IEEE, 2024.

Abstract

In recent years, transformer language models have made a significant impact on automatic text generation. This study focuses on the task of distractor generation in Spanish using a fine-tuned multilingual text-to-text model, namely mT5. Our method outperformed established baselines based on LSTM networks, confirming the effectiveness of Transformer architectures in such NLP tasks. While comparisons with other Transformer-based solutions yielded diverse outcomes based on the metric of choice, our method notably achieved superior results on the ROUGE metric compared to the GPT-2 approach. Although traditional evaluation metrics such as BLEU and ROUGE are commonly used, this paper argues for more context-sensitive metrics given the inherent variability in acceptable distractor generation results. Among the contributions of this research is a comprehensive comparison with other methods, an examination of the potential drawbacks of multilingual models, and the introduction of alternative evaluation metrics. Future research directions, derived from our findings and a review of related works are also suggested, with a particular emphasis on leveraging other language models and Transformer architectures.

Details

Language :
English
ISSN :
21693536
Volume :
12
Database :
Directory of Open Access Journals
Journal :
IEEE Access
Publication Type :
Academic Journal
Accession number :
edsdoj.9d0e27c3295a4d948b6557d910b321ff
Document Type :
article
Full Text :
https://doi.org/10.1109/ACCESS.2024.3361673