1. Application of Quantum Recurrent Neural Network in Low-Resource Language Text Classification
- Author
-
Wenbin Yu, Lei Yin, Chengjun Zhang, Yadang Chen, and Alex X. Liu
- Subjects
Natural language processing (NLP) ,quantum machine learning ,quantum recurrent neural network ,Atomic physics. Constitution and properties of matter ,QC170-197 ,Materials of engineering and construction. Mechanics of materials ,TA401-492 - Abstract
Text sentiment analysis is an important task in natural language processing and has always been a hot research topic. However, in low-resource regions such as South Asia, where languages like Bengali are widely used, the research interest is relatively low compared to high-resource regions due to limited computational resources, flexible word order, and high inflectional nature of the language. With the development of quantum technology, quantum machine learning models leverage the superposition property of qubits to enhance model expressiveness and achieve faster computation compared to classical systems. To promote the development of quantum machine learning in low-resource language domains, we propose a quantum–classical hybrid architecture. This architecture utilizes a pretrained multilingual bidirectional encoder representations from transformer (BERT) model to obtain vector representations of words and combines the proposed batch upload quantum recurrent neural network (BUQRNN) and parameter nonshared batch upload quantum recurrent neural network (PN-BUQRNN) as feature extraction models for sentiment analysis in Bengali. Our numerical results demonstrate that the proposed BUQRNN structure achieves a maximum accuracy improvement of 0.993% in Bengali text classification tasks while reducing average model complexity by 12%. The PN-BUQRNN structure surpasses the BUQRNN structure once again and outperforms classical architectures in certain tasks.
- Published
- 2024
- Full Text
- View/download PDF