Back to Search Start Over

Robust Chinese Short Text Entity Disambiguation Method Based on Feature Fusion and Contrastive Learning.

Authors :
Mei, Qishun
Li, Xuhui
Source :
Information (2078-2489). Mar2024, Vol. 15 Issue 3, p139. 17p.
Publication Year :
2024

Abstract

To address the limitations of existing methods of short-text entity disambiguation, specifically in terms of their insufficient feature extraction and reliance on massive training samples, we propose an entity disambiguation model called COLBERT, which fuses LDA-based topic features and BERT-based semantic features, as well as using contrastive learning, to enhance the disambiguation process. Experiments on a publicly available Chinese short-text entity disambiguation dataset show that the proposed model achieves an F1-score of 84.0%, which outperforms the benchmark method by 0.6%. Moreover, our model achieves an F1-score of 74.5% with a limited number of training samples, which is 2.8% higher than the benchmark method. These results demonstrate that our model achieves better effectiveness and robustness and can reduce the burden of data annotation as well as training costs. [ABSTRACT FROM AUTHOR]

Subjects

Subjects :
*FEATURE extraction
*ANNOTATIONS

Details

Language :
English
ISSN :
20782489
Volume :
15
Issue :
3
Database :
Academic Search Index
Journal :
Information (2078-2489)
Publication Type :
Academic Journal
Accession number :
176334012
Full Text :
https://doi.org/10.3390/info15030139