Back to Search Start Over

Utilizing Recurrent Neural Network for topic discovery in short text scenarios1.

Authors :
Lu, Heng-Yang
Kang, Ning
Li, Yun
Zhan, Qian-Yi
Xie, Jun-Yuan
Wang, Chong-Jun
Source :
Intelligent Data Analysis. 2019, Vol. 23 Issue 2, p259-277. 19p.
Publication Year :
2019

Abstract

The volume of short text data increases rapidly these years. Data examples include tweets and online Q&A pairs. It is essential to organize and summarize these data automatically. Topic model is one of the effective approaches, whose application domains include text mining, personalized recommendation and so on. Conventional models like pLSA and LDA are designed for long text data. However, these models may suffer from the sparsity problem brought by lacking words in short text scenarios. Recent studies such as BTM show that using word co-occurrent pairs is effective to relieve the sparsity problem. However, both BTM and extended models ignore the quantifiable relationship between words. From our perspectives, two more related words should occur in the same topic. Based on this idea, we introduce a model named RIBS, which makes use of RNN to learn relationship. By using the learned relationship, we introduce a model named RIBS-Bigrams, which can display topics with bigrams. Through experiments on two open-source and real-world datasets, RIBS achieves better coherence in topic discovery, and RIBS-Bigrams achieves better readability in topic display. In the document characterization task, the document representation of RIBS can lead better purity and entropy in clustering, higher accuracy in classification. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
1088467X
Volume :
23
Issue :
2
Database :
Academic Search Index
Journal :
Intelligent Data Analysis
Publication Type :
Academic Journal
Accession number :
135826007
Full Text :
https://doi.org/10.3233/IDA-183842