Back to Search Start Over

Comparative Study of Multiclass Text Classification in Research Proposals Using Pretrained Language Models

Authors :
Eunchan Lee
Changhyeon Lee
Sangtae Ahn
Source :
Applied Sciences; Volume 12; Issue 9; Pages: 4522
Publication Year :
2022
Publisher :
Multidisciplinary Digital Publishing Institute, 2022.

Abstract

Recently, transformer-based pretrained language models have demonstrated stellar performance in natural language understanding (NLU) tasks. For example, bidirectional encoder representations from transformers (BERT) have achieved outstanding performance through masked self-supervised pretraining and transformer-based modeling. However, the original BERT may only be effective for English-based NLU tasks, whereas its effectiveness for other languages such as Korean is limited. Thus, the applicability of BERT-based language models pretrained in languages other than English to NLU tasks based on those languages must be investigated. In this study, we comparatively evaluated seven BERT-based pretrained language models and their expected applicability to Korean NLU tasks. We used the climate technology dataset, which is a Korean-based large text classification dataset, in research proposals involving 45 classes. We found that the BERT-based model pretrained on the most recent Korean corpus performed the best in terms of Korean-based multiclass text classification. This suggests the necessity of optimal pretraining for specific NLU tasks, particularly those in languages other than English.

Details

Language :
English
ISSN :
20763417
Database :
OpenAIRE
Journal :
Applied Sciences; Volume 12; Issue 9; Pages: 4522
Accession number :
edsair.doi.dedup.....1546b5e343fe541571efb6b5c46265f4
Full Text :
https://doi.org/10.3390/app12094522