计算机科学
模式
模态(人机交互)
人工智能
情态动词
特征学习
特征(语言学)
代表(政治)
判决
自然语言处理
情绪分析
语义学(计算机科学)
深度学习
机器学习
模式识别(心理学)
语言学
社会科学
化学
哲学
社会学
政治
政治学
高分子化学
法学
程序设计语言
作者
Mi-Xiao Hou,Zheng Zhang,Guangming Lu
标识
DOI:10.1109/icassp43922.2022.9747859
摘要
Multi-modal emotion recognition aims to extract sentiment-related information from multiple sources and integrate different modal representations for sentiment analysis. Alignment is an effective strategy to achieve semantically consistent representations for multi-modal emotion recognition, while the current alignment models are jointly unable to maintain the dependence of word-to-sentence and independence of unimodal learning. In this paper, we propose a Self-guided Modality Calibration Network (SMCN) to realize multi-modal alignment which can capture the global connections without interfering with unimodal learning. While preserving unimodal learning without interference, our model leverages semantic sentiment-related features to guide modality-specific representation learning. On one hand, SMCN simulates human thinking by deriving a branch for acquiring knowledge of other modalities in unimodal learning. This branch aims to lean high-level semantic information of other modalities for realizing semantic alignment between modalities. On the other hand, we also provide an indirect interaction manner to integrate unimodal feature and calibrate features in different levels for avoiding unimodal features mixed with other clues. Experiments demonstrate that our approach outperforms the state-of-the-art methods on both IEMOCAP and MELD datasets.
科研通智能强力驱动
Strongly Powered by AbleSci AI