Back to Search Start Over

Cross-linguistic comparison of linguistic feature encoding in BERT models for typologically different languages

Authors :
Otmakhova, Y
Verspoor, K
Lau, JH
Otmakhova, Y
Verspoor, K
Lau, JH
Source :
4th Workshop on Research in Computational Linguistic Typology and Multilingual NLP
Publication Year :
2022

Abstract

Though recently there have been an increased interest in how pre-trained language models encode different linguistic features, there is still a lack of systematic comparison between languages with different morphology and syntax. In this paper, using BERT as an example of a pre-trained model, we compare how three typologically different languages (English, Korean, and Russian) encode morphology and syntax features across different layers. In particular, we contrast languages which differ in a particular aspect, such as flexibility of word order, head directionality, morphological type, presence of grammatical gender, and morphological richness, across four different tasks.

Details

Database :
OAIster
Journal :
4th Workshop on Research in Computational Linguistic Typology and Multilingual NLP
Publication Type :
Electronic Resource
Accession number :
edsoai.on1373007612
Document Type :
Electronic Resource