Back to Search
Start Over
Robust Chinese Short Text Entity Disambiguation Method Based on Feature Fusion and Contrastive Learning.
- Source :
-
Information (2078-2489) . Mar2024, Vol. 15 Issue 3, p139. 17p. - Publication Year :
- 2024
-
Abstract
- To address the limitations of existing methods of short-text entity disambiguation, specifically in terms of their insufficient feature extraction and reliance on massive training samples, we propose an entity disambiguation model called COLBERT, which fuses LDA-based topic features and BERT-based semantic features, as well as using contrastive learning, to enhance the disambiguation process. Experiments on a publicly available Chinese short-text entity disambiguation dataset show that the proposed model achieves an F1-score of 84.0%, which outperforms the benchmark method by 0.6%. Moreover, our model achieves an F1-score of 74.5% with a limited number of training samples, which is 2.8% higher than the benchmark method. These results demonstrate that our model achieves better effectiveness and robustness and can reduce the burden of data annotation as well as training costs. [ABSTRACT FROM AUTHOR]
- Subjects :
- *FEATURE extraction
*ANNOTATIONS
Subjects
Details
- Language :
- English
- ISSN :
- 20782489
- Volume :
- 15
- Issue :
- 3
- Database :
- Academic Search Index
- Journal :
- Information (2078-2489)
- Publication Type :
- Academic Journal
- Accession number :
- 176334012
- Full Text :
- https://doi.org/10.3390/info15030139