Diversity and Balance: Multimodal Sentiment Analysis Using Multimodal-Prefixed and Cross-Modal Attention

情态动词 平衡(能力) 多样性(政治) 情绪分析 多模态 模式治疗法 计算机科学 心理学 语言学 认知心理学 人工智能 社会学 哲学 人类学 神经科学 化学 高分子化学 心理治疗师
作者
Meng Li,Zhenfang Zhu,Kefeng Li,Hongli Pei
出处
期刊:IEEE Transactions on Affective Computing [Institute of Electrical and Electronics Engineers]
卷期号:16 (1): 250-263 被引量:14
标识
DOI:10.1109/taffc.2024.3430045
摘要

Multimodal Sentiment Analysis (MSA) is the technology of intelligently recognizing and assessing human sentiments using various data forms such as text, image, and audio. Despite current mainstream methods have made significant progress, MSA still faces the following issues: 1) most current methods train models based on pre-extracted features, lacking a sufficient understanding of sentiment diversity in multimodal data and may even lead to the loss of critical information in the raw data; and 2) textual modality, which possesses high-level semantic features, should typically dominate the fusion process, yet current methods fail to fully leverage this characteristic to balance modality information. To address the aforementioned issues, we propose a novel Multimodal Sentiment Analysis framework using Multimodal-Prefixed and Cross-Modal Attention (DB-MPCA). For the first issue, DB-MPCA employs multimodal raw data for pre-training, which not only allows for in-depth exploration of multimodal information but also significantly enhances the model’s learning capabilities and generalization, while reducing the substantial costs associated with manual annotation. Regarding the second issue, DB-MPCA introduces two prefix encoders designed to convert acoustic and visual features into prefix tokens. These tokens are then embedded into a pre-trained language model, where they are encoded together with textual tokens. Through this approach, DB-MPCA effectively learns cross-modal attention while maintaining the dominance of the textual modality, thereby optimizing the fusion of modalities. Comprehensive experiments conducted on the widely utilized dataset (CMU-MOSI) demonstrate the effectiveness of our model, highlighting its superiority over baseline models.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
脑洞疼应助成就的绯采纳,获得10
1秒前
SciGPT应助成就的井采纳,获得10
2秒前
佰慧完成签到,获得积分10
2秒前
2秒前
myyyyy完成签到,获得积分10
4秒前
4秒前
叶子麻完成签到,获得积分10
4秒前
4秒前
6秒前
成熟稳重痴情完成签到,获得积分10
6秒前
ppsweek发布了新的文献求助10
6秒前
慕青应助cgjhgh采纳,获得10
7秒前
英姑应助Lww采纳,获得10
7秒前
9秒前
搜集达人应助你好采纳,获得10
9秒前
英俊的铭应助ppsweek采纳,获得10
10秒前
逗我呢发布了新的文献求助10
10秒前
11秒前
liriyii发布了新的文献求助10
11秒前
成就的绯完成签到,获得积分10
13秒前
卡尔完成签到,获得积分10
14秒前
俭朴觅松完成签到 ,获得积分10
15秒前
张张完成签到,获得积分10
15秒前
xsaltycafe发布了新的文献求助10
15秒前
17秒前
聪慧的凝海完成签到 ,获得积分0
18秒前
莎莎完成签到 ,获得积分10
18秒前
研友_VZG7GZ应助油柑美式采纳,获得20
19秒前
叶保听完成签到,获得积分10
19秒前
19秒前
兰亭序完成签到 ,获得积分10
19秒前
云馨完成签到,获得积分10
20秒前
20秒前
NexusExplorer应助Glacier采纳,获得10
23秒前
CC发布了新的文献求助10
23秒前
24秒前
你好发布了新的文献求助10
26秒前
jlw完成签到,获得积分10
26秒前
28秒前
许平平完成签到,获得积分20
28秒前
高分求助中
(应助此贴封号)【重要!!请各用户(尤其是新用户)详细阅读】【科研通的精品贴汇总】 10000
The Cambridge History of China: Volume 4, Sui and T'ang China, 589–906 AD, Part Two 1500
Cowries - A Guide to the Gastropod Family Cypraeidae 1200
Quality by Design - An Indispensable Approach to Accelerate Biopharmaceutical Product Development 800
Signals, Systems, and Signal Processing 610
Research Methods for Applied Linguistics 500
A Social and Cultural History of the Hellenistic World 500
热门求助领域 (近24小时)
化学 材料科学 医学 生物 纳米技术 工程类 有机化学 化学工程 生物化学 计算机科学 物理 内科学 复合材料 催化作用 物理化学 光电子学 电极 细胞生物学 基因 无机化学
热门帖子
关注 科研通微信公众号,转发送积分 6395884
求助须知:如何正确求助?哪些是违规求助? 8211281
关于积分的说明 17392783
捐赠科研通 5449406
什么是DOI,文献DOI怎么找? 2880466
邀请新用户注册赠送积分活动 1857078
关于科研通互助平台的介绍 1699428