Back to Search Start Over

An autoencoder-based self-supervised learning for multimodal sentiment analysis.

Authors :
Feng, Wenjun
Wang, Xin
Cao, Donglin
Lin, Dazhen
Source :
Information Sciences. Jul2024, Vol. 675, pN.PAG-N.PAG. 1p.
Publication Year :
2024

Abstract

Representation learning is a crucial and challenging task within multimodal sentiment analysis. Effective multimodal sentiment representations contain two key aspects: consistency and difference. However, the state-of-the-art multimodal sentiment analysis approaches failed to capture the difference and consistency of sentiment information across diverse modalities. To address the multimodal sentiment representation problem, we propose an autoencoder-based self-supervised learning framework. In the pre-training stage, an autoencoder is designed for each modality, leveraging unlabeled data to learn richer sentiment representations for each modality through sample reconstruction and modality consistency detection tasks. In the fine-tuning stage, the pre-trained autoencoder is injected into MulT (AE-MT) and enhance the model's ability to extract deep sentiment information by incorporating a contrastive learning auxiliary task. Our experiments on the popular Chinese sentiment analysis benchmark (CH-SIMS v2.0) and English sentiment analysis benchmark (MOSEI) demonstrate significant gains over baseline models. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
00200255
Volume :
675
Database :
Academic Search Index
Journal :
Information Sciences
Publication Type :
Periodical
Accession number :
177602245
Full Text :
https://doi.org/10.1016/j.ins.2024.120682