Back to Search Start Over

A generative summarization model combined NLG and NLU.

Authors :
Lv, Fangxing
Liu, Wenfeng
Yang, Yuzhen
Gao, Yaling
Bao, Longqing
Source :
Journal of Intelligent & Fuzzy Systems. Apr2024, p1-9. 9p.
Publication Year :
2024

Abstract

The automatic generation of natural language is a complex and essential task in text processing. This study proposes a novel approach to address this fundamental problem by leveraging an improved version of DST_BERT, a model that converts input text into a vector representation. Our key contribution lies in the joint optimization of two models, <italic>NLU</italic> (Natural Language Under-standing) and <italic>NLG</italic> (Natural Language Generation), which enables us to obtain variable representations within a hidden space. This integration enhances the capabilities of both <italic>NLU</italic> and <italic>NLG</italic> in generating coherent and contextually appropriate language. The <italic>NLU</italic> and <italic>NLG</italic> models are seamlessly integrated with the hidden variable space, forming a generative representation model. To assess the effectiveness of our proposed approach, we conducted extensive experiments on the E2E and Weather datasets. The results highlight the state-of-the-art performance achieved by our model in generating natural language. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
10641246
Database :
Academic Search Index
Journal :
Journal of Intelligent & Fuzzy Systems
Publication Type :
Academic Journal
Accession number :
176672439
Full Text :
https://doi.org/10.3233/jifs-232981