Back to Search
Start Over
Wikipedia Entities as Rendezvous across Languages: Grounding Multilingual Language Models by Predicting Wikipedia Hyperlinks
- Source :
- The 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: NAACL-HLT 2021 : proceedings of the conference : June 6-11, 2021, 3651-3661, STARTPAGE=3651;ENDPAGE=3661;TITLE=The 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT, Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
- Publication Year :
- 2021
- Publisher :
- The Association for Computational Linguistics, 2021.
-
Abstract
- Masked language models have quickly become the de facto standard when processing text. Recently, several approaches have been proposed to further enrich word representations with external knowledge sources such as knowledge graphs. However, these models are devised and evaluated in a monolingual setting only. In this work, we propose a language-independent entity prediction task as an intermediate training procedure to ground word representations on entity semantics and bridge the gap across different languages by means of a shared vocabulary of entities. We show that our approach effectively injects new lexical-semantic knowledge into neural models, improving their performance on different semantic tasks in the zero-shot crosslingual setting. As an additional advantage, our intermediate training does not require any supplementary input, allowing our models to be applied to new datasets right away. In our experiments, we use Wikipedia articles in up to 100 languages and already observe consistent gains compared to strong baselines when predicting entities using only the English Wikipedia. Further adding extra languages lead to improvements in most tasks up to a certain point, but overall we found it non-trivial to scale improvements in model transferability by training on ever increasing amounts of Wikipedia languages.
- Subjects :
- Vocabulary
Computer science
media_common.quotation_subject
02 engineering and technology
entity linking
computer.software_genre
Semantics
03 medical and health sciences
0302 clinical medicine
0202 electrical engineering, electronic engineering, information engineering
6121 Languages
media_common
Point (typography)
business.industry
deep learning
Hyperlink
113 Computer and information sciences
Task (computing)
030221 ophthalmology & optometry
transformer
language model
020201 artificial intelligence & image processing
Language model
Artificial intelligence
multilingual
business
computer
Word (computer architecture)
Natural language processing
De facto standard
Subjects
Details
- Language :
- English
- Database :
- OpenAIRE
- Journal :
- The 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: NAACL-HLT 2021 : proceedings of the conference : June 6-11, 2021, 3651-3661, STARTPAGE=3651;ENDPAGE=3661;TITLE=The 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT, Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
- Accession number :
- edsair.doi.dedup.....ab4d34426d9da2cc48fcdaafa8e51756