Back to Search
Start Over
MG-BERT: Multi-Graph Augmented BERT for Masked Language Modeling
- Source :
- Proceedings of the Fifteenth Workshop on Graph-Based Methods for Natural Language Processing (TextGraphs-15).
- Publication Year :
- 2021
- Publisher :
- Association for Computational Linguistics, 2021.
-
Abstract
- Pre-trained models like Bidirectional Encoder Representations from Transformers (BERT), have recently made a big leap forward in Natural Language Processing (NLP) tasks. However, there are still some shortcomings in the Masked Language Modeling (MLM) task performed by these models. In this paper, we first introduce a multi-graph including different types of relations between words. Then, we propose Multi-Graph augmented BERT (MG-BERT) model that is based on BERT. MG-BERT embeds tokens while taking advantage of a static multi-graph containing global word co-occurrences in the text corpus beside global real-world facts about words in knowledge graphs. The proposed model also employs a dynamic sentence graph to capture local context effectively. Experimental results demonstrate that our model can considerably enhance the performance in the MLM task.
- Subjects :
- Text corpus
business.industry
Computer science
Context (language use)
02 engineering and technology
computer.software_genre
Task (project management)
03 medical and health sciences
0302 clinical medicine
030221 ophthalmology & optometry
0202 electrical engineering, electronic engineering, information engineering
Graph (abstract data type)
020201 artificial intelligence & image processing
Artificial intelligence
Language model
business
Encoder
computer
Sentence
Word (computer architecture)
Natural language processing
Subjects
Details
- Database :
- OpenAIRE
- Journal :
- Proceedings of the Fifteenth Workshop on Graph-Based Methods for Natural Language Processing (TextGraphs-15)
- Accession number :
- edsair.doi.dedup.....3dc85ee250c9f53fdb6eaf279172374f
- Full Text :
- https://doi.org/10.18653/v1/11.textgraphs-1.12