Back to Search Start Over

Constructing Chinese taxonomy trees from understanding and generative pretrained language models

Authors :
Jianyu Guo
Jingnan Chen
Li Ren
Huanlai Zhou
Wenbo Xu
Haitao Jia
Source :
PeerJ Computer Science, Vol 10, p e2358 (2024)
Publication Year :
2024
Publisher :
PeerJ Inc., 2024.

Abstract

The construction of hypernym taxonomic trees, a critical task in the field of natural language processing, involves extracting lexical relationships, specifically creating a tree structure that represents hypernym relationships among a given set of words within the same domain. In this work, we present a method for constructing hypernym taxonomy trees in the Chinese language domain, and we named it CHRRM (Chinese Hypernym Relationship Reasoning Model). Our method consists of two main steps: First, we utilize pre-trained models to predict hypernym relationships between pairs of words; second, we regard these relationships as edges to form a maximum spanning tree in the word graph. Our method enhances the effectiveness of constructing hypernym taxonomic trees based on pre-trained models through two key improvements: (1) We optimize the hyperparameter configuration for this task using pre-trained models from the Bert family and provide explanations for the configuration of these hyperparameters. (2) By employing generative large language models such as ChatGPT and ChatGLM to annotate words, we improve the accuracy of hypernym relationship identification and analyze the feasibility of applying generative large language models to the task of constructing taxonomy trees. We trained our model on subtrees of WORDNET and evaluated its performance on non-overlapping subtrees of WORDNET, demonstrating that our enhancements led to a significant relative improvement of 15.67%, achieving an F1 score of 67.9 on the Chinese WORDNET validation dataset compared to the previous score of 58.7. In conclusion, our study reveals the following key findings: (1) The Roberta-wwm-ext-large model consistently delivers outstanding results in constructing taxonomic trees. (2) Generative large language models, while capable of aiding pre-trained models in improving hypernym recognition accuracy, have limitations related to generation quality and computational resources. (3) Generative large language models can serve various NLP tasks either directly or indirectly; it is feasible to improve the downstream NLU task’s performance through the generative content.

Details

Language :
English
ISSN :
23765992
Volume :
10
Database :
Directory of Open Access Journals
Journal :
PeerJ Computer Science
Publication Type :
Academic Journal
Accession number :
edsdoj.ba3e85af1679408981bef0b39d4f2e9e
Document Type :
article
Full Text :
https://doi.org/10.7717/peerj-cs.2358