计算机科学
人工智能
模态(人机交互)
自然语言处理
情绪分析
组分(热力学)
特征(语言学)
代表(政治)
情态动词
特征学习
模式识别(心理学)
特征向量
语言学
哲学
化学
高分子化学
法学
物理
热力学
政治
政治学
作者
Xiaocui Yang,Shi Feng,Daling Wang,Pengfei Hong,Soujanya Poria
标识
DOI:10.1109/icassp49357.2023.10096777
摘要
Multimodal sentiment analysis has received extensive attention with the explosion of multimodal data. For multimodal data, representations should have disparate distributions in the feature space under different labels. The paired multi-modal image-text posts should be closer than unpaired. We propose Multimodal fine-grained interaction with the Multiple Contrastive Learning (M 2 CL) model for image-text multi-modal sentiment detection. Specifically, we first obtain the reinforced global representation of one modality with the assistance of fine-grained information from another via the Multimodal Interaction Component. Then, we introduce the Multiple Contrastive Learning Component, including Supervised Contrastive Learning (SCL) and Dual Multimodal Contrastive Learning (DMCL). SCL accomplishes pushing the posts with the same sentiment closer and pulling the instances of different sentiments apart within each modality. DMCL pushes the paired image-text features together and pulls the unpaired apart with multiple stages. Extensive experiments conducted on three datasets confirm the effectiveness of our approach.
科研通智能强力驱动
Strongly Powered by AbleSci AI