Back to Search
Start Over
Cross-linguistic comparison of linguistic feature encoding in BERT models for typologically different languages
- Source :
- 4th Workshop on Research in Computational Linguistic Typology and Multilingual NLP
- Publication Year :
- 2022
-
Abstract
- Though recently there have been an increased interest in how pre-trained language models encode different linguistic features, there is still a lack of systematic comparison between languages with different morphology and syntax. In this paper, using BERT as an example of a pre-trained model, we compare how three typologically different languages (English, Korean, and Russian) encode morphology and syntax features across different layers. In particular, we contrast languages which differ in a particular aspect, such as flexibility of word order, head directionality, morphological type, presence of grammatical gender, and morphological richness, across four different tasks.
Details
- Database :
- OAIster
- Journal :
- 4th Workshop on Research in Computational Linguistic Typology and Multilingual NLP
- Publication Type :
- Electronic Resource
- Accession number :
- edsoai.on1373007612
- Document Type :
- Electronic Resource