Back to Search
Start Over
MASANet: Multi-Aspect Semantic Auxiliary Network for Visual Sentiment Analysis.
- Source :
- IEEE Transactions on Affective Computing; Jul-Sep2024, Vol. 15 Issue 3, p1439-1450, 12p
- Publication Year :
- 2024
-
Abstract
- Recently, multi-modal affective computing has demonstrated that introducing multi-modal information can enhance performance. However, multi-modal research faces significant challenges due to its high requirements regarding data acquisition, modal integrity, and feature alignment. The widespread use of multi-modal pre-training methods offers the possibility of aiding visual sentiment analysis by introducing cross-domain knowledge. This paper proposes a Multi-Aspect Semantic Auxiliary Network (MASANet) for visual sentiment analysis. Specifically, MASANet achieves modality expansion through cross-modal generation, making it possible to introduce cross-domain semantic assistance. Then, a cross-modal gating module and an adaptive modal fusion module are proposed for aspect-level and cross-modal interaction, respectively. In addition, a designed semantic polarity constraint loss is presented to improve sentiment multi-classification performance. Evaluations of eight widely-used affective image datasets demonstrate that our proposed method outperforms the state-of-the-art methods. Further ablation experiments and visualization results also confirm the effectiveness of the proposed method and its modules. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 19493045
- Volume :
- 15
- Issue :
- 3
- Database :
- Complementary Index
- Journal :
- IEEE Transactions on Affective Computing
- Publication Type :
- Academic Journal
- Accession number :
- 179509544
- Full Text :
- https://doi.org/10.1109/TAFFC.2023.3331776