情绪分析
情态动词
翻译(生物学)
计算机科学
特征(语言学)
人工智能
自然语言处理
语音识别
模式识别(心理学)
语言学
材料科学
化学
生物化学
哲学
信使核糖核酸
高分子化学
基因
作者
Chenquan Gan,Yu Tang,Xiang Fu,Qingyi Zhu,Deepak Kumar Jain,Salvador García
标识
DOI:10.1016/j.knosys.2024.111982
摘要
Multimodal sentiment analysis on social platforms is crucial for comprehending public opinions and attitudes, thus garnering substantial interest in knowledge engineering. Existing methods like implicit interaction, explicit interaction, and cross-modal translation can effectively integrate sentiment information, but they encounter challenges in establishing efficient emotional correlations across modalities due to data heterogeneity and concealed emotional relationships. To tackle this issue, we propose a video multimodal sentiment analysis model called PEST, which leverages cross-modal feature translation and a dynamic propagation model. Specifically, cross-modal feature translation translates textual, visual, and acoustic features into a common feature space, eliminating heterogeneity and enabling initial modal interaction. Additionally, the dynamic propagation model facilitates in-depth interaction and aids in establishing stable and reliable emotional correlations across modalities. Extensive experiments on the three multimodal sentiment datasets, CMU-MOSI, CMU-MOSEI, and CH-SIMS, demonstrate that PEST exhibits superior performance in both word-aligned and unaligned settings.
科研通智能强力驱动
Strongly Powered by AbleSci AI