1. Answering Fill-in-the-Blank Questions in Portuguese with Transformer Language Models
- Author
-
Hugo Gonçalo Oliveira
- Subjects
Sequence ,business.industry ,Semantics (computer science) ,Computer science ,computer.software_genre ,Blank ,language.human_language ,language ,Language modelling ,Language model ,Artificial intelligence ,Portuguese ,business ,computer ,Natural language processing ,Sentence ,Transformer (machine learning model) - Abstract
Despite different applications, transformer-based language models, like BERT and GPT, learn about language by predicting missing parts of text. BERT is pretrained in Masked Language Modelling and GPT generates text from a given sequence. We explore such models for answering cloze questions in Portuguese, following different approaches. When options are not considered, the largest BERT model, trained exclusively for Portuguese, is the most accurate. But when selecting the best option, top performance is achieved by computing the most probable sentence, and GPT-2 fine-tuned for Portuguese beats BERT.
- Published
- 2021
- Full Text
- View/download PDF