1. Label-Embedding Bi-directional Attentive Model for Multi-label Text Classification
- Author
-
Jiangtao Ren, Naiyin Liu, and Qianlong Wang
- Subjects
0209 industrial biotechnology ,Language representation ,Computer Networks and Communications ,business.industry ,Computer science ,General Neuroscience ,Computational intelligence ,02 engineering and technology ,computer.software_genre ,Security token ,Field (computer science) ,Task (project management) ,ComputingMethodologies_PATTERNRECOGNITION ,020901 industrial engineering & automation ,Artificial Intelligence ,0202 electrical engineering, electronic engineering, information engineering ,Embedding ,020201 artificial intelligence & image processing ,Artificial intelligence ,State (computer science) ,business ,Representation (mathematics) ,computer ,Software ,Natural language processing - Abstract
Multi-label text classification is a critical task in natural language processing field. As the latest language representation model, BERT obtains new state-of-the-art results in the classification task. Nevertheless, the text classification framework of BERT neglects to make full use of the token-level text representation and label embedding, since it only utilizes the final hidden state corresponding to CLS token as sequence-level text representation for classification. We assume that the finer-grained token-level text representation and label embedding contribute to classification. Consequently, in this paper, we propose a Label-Embedding Bi-directional Attentive model to improve the performance of BERT’s text classification framework. In particular, we extend BERT’s text classification framework with label embedding and bi-directional attention. Experimental results on the five datasets indicate that our model has notable improvements over both baselines and state-of-the-art models.
- Published
- 2021
- Full Text
- View/download PDF