计算机科学
情态动词
人工智能
情绪分析
可视化
领域(数学分析)
约束(计算机辅助设计)
特征(语言学)
模式识别(心理学)
机器学习
自然语言处理
工程类
数学分析
哲学
机械工程
化学
高分子化学
语言学
数学
作者
Jinglun Cen,Chunmei Qing,Haochun Ou,Xiangmin Xu,Junpeng Tan
标识
DOI:10.1109/taffc.2023.3331776
摘要
Recently, multi-modal affective computing has demonstrated that introducing multi-modal information can enhance performance. However, multi-modal research faces significant challenges due to its high requirements regarding data acquisition, modal integrity, and feature alignment. The widespread use of multi-modal pre-training methods offers the possibility of aiding visual sentiment analysis by introducing cross-domain knowledge. This paper proposes a Multi-Aspect Semantic Auxiliary Network (MASANet) for visual sentiment analysis. Specifically, MASANet achieves modality expansion through cross-modal generation, making it possible to introduce cross-domain semantic assistance. Then, a cross-modal gating module and an adaptive modal fusion module are proposed for aspect-level and cross-modal interaction, respectively. In addition, a designed semantic polarity constraint loss is presented to improve sentiment multi-classification performance. Evaluations of eight widely-used affective image datasets demonstrate that our proposed method outperforms the state-of-the-art methods. Further ablation experiments and visualization results also confirm the effectiveness of the proposed method and its modules.
科研通智能强力驱动
Strongly Powered by AbleSci AI