Back to Search Start Over

Deep learning-based multimodal emotion recognition from audio, visual, and text modalities: A systematic review of recent advancements and future prospects.

Authors :
Zhang, Shiqing
Yang, Yijiao
Chen, Chen
Zhang, Xingnan
Leng, Qingming
Zhao, Xiaoming
Source :
Expert Systems with Applications. Mar2024:Part C, Vol. 237, pN.PAG-N.PAG. 1p.
Publication Year :
2024

Abstract

Emotion recognition has recently attracted extensive interest due to its significant applications to human–computer interaction. The expression of human emotion depends on various verbal and non-verbal languages like audio, visual, text, etc. Emotion recognition is thus well suited as a multimodal rather than single-modal learning problem. Owing to the powerful feature learning capability, extensive deep learning methods have been recently leveraged to capture high-level emotional feature representations for multimodal emotion recognition (MER). Therefore, this paper makes the first effort in comprehensively summarize recent advances in deep learning-based multimodal emotion recognition (DL-MER) involved in audio, visual, and text modalities. We focus on: (1) MER milestones are given to summarize the development tendency of MER, and conventional multimodal emotional datasets are provided; (2) The core principles of typical deep learning models and its recent advancements are overviewed; (3) A systematic survey and taxonomy is provided to cover the state-of-the-art methods related to two key steps in a MER system, including feature extraction and multimodal information fusion; (4) The research challenges and open issues in this field are discussed, and promising future directions are given. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
09574174
Volume :
237
Database :
Academic Search Index
Journal :
Expert Systems with Applications
Publication Type :
Academic Journal
Accession number :
173631582
Full Text :
https://doi.org/10.1016/j.eswa.2023.121692