亲爱的研友该休息了!由于当前在线用户较少,发布求助请尽量完整地填写文献信息,科研通机器人24小时在线,伴您度过漫漫科研夜!身体可是革命的本钱,早点休息,好梦!

Multimodal Emotion Recognition Fusion Analysis Adapting BERT With Heterogeneous Feature Unification

计算机科学 模式 人工智能 情绪分析 语音识别 情感计算 特征(语言学) 模态(人机交互) 面部表情 多模式学习 人机交互 社会科学 语言学 哲学 社会学
作者
SangHyun Lee,David K. Han,Hanseok Ko
出处
期刊:IEEE Access [Institute of Electrical and Electronics Engineers]
卷期号:9: 94557-94572 被引量:3
标识
DOI:10.1109/access.2021.3092735
摘要

Human communication includes rich emotional content, thus the development of multimodal emotion recognition plays an important role in communication between humans and computers. Because of the complex emotional characteristics of a speaker, emotional recognition remains a challenge, particularly in capturing emotional cues across a variety of modalities, such as speech, facial expressions, and language. Audio and visual cues are particularly vital for a human observer in understanding emotions. However, most previous work on emotion recognition has been based solely on linguistic information, which can overlook various forms of nonverbal information. In this paper, we present a new multimodal emotion recognition approach that improves the BERT model for emotion recognition by combining it with heterogeneous features based on language, audio, and visual modalities. Specifically, we improve the BERT model due to the heterogeneous features of the audio and visual modalities. We introduce the Self-Multi-Attention Fusion module, Multi-Attention fusion module, and Video Fusion module, which are attention based multimodal fusion mechanisms using the recently proposed transformer architecture. We explore the optimal ways to combine fine-grained representations of audio and visual features into a common embedding while combining a pre-trained BERT model with modalities for fine-tuning. In our experiment, we evaluate the commonly used CMU-MOSI, CMU-MOSEI, and IEMOCAP datasets for multimodal sentiment analysis. Ablation analysis indicates that the audio and visual components make a significant contribution to the recognition results, suggesting that these modalities contain highly complementary information for sentiment analysis based on video input. Our method shows that we achieve state-of-the-art performance on the CMU-MOSI, CMU-MOSEI, and IEMOCAP dataset.

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
香蕉觅云应助stupidZ采纳,获得30
刚刚
细腻的钢铁侠完成签到,获得积分10
刚刚
1秒前
英姑应助友好灵阳采纳,获得10
3秒前
余杭发布了新的文献求助10
5秒前
6秒前
7秒前
打打应助余杭采纳,获得10
11秒前
Yuuuu发布了新的文献求助10
12秒前
胡大笑哈哈哈完成签到 ,获得积分10
12秒前
13秒前
13秒前
13秒前
Hu发布了新的文献求助100
15秒前
16秒前
18秒前
欣阳1021发布了新的文献求助10
18秒前
19秒前
scfsl发布了新的文献求助10
20秒前
20秒前
小华完成签到 ,获得积分10
22秒前
Gates发布了新的文献求助10
22秒前
yanglinhai完成签到 ,获得积分10
23秒前
屈春洋发布了新的文献求助10
24秒前
852应助傲娇的小松鼠采纳,获得10
25秒前
25秒前
木华驳回了华仔应助
26秒前
充电宝应助Gates采纳,获得10
27秒前
MOFS完成签到,获得积分10
29秒前
29秒前
32秒前
电量过低完成签到 ,获得积分10
33秒前
39秒前
43秒前
44秒前
清秀小霸王完成签到 ,获得积分10
45秒前
丘比特应助komorebi采纳,获得10
47秒前
斯文的白玉完成签到,获得积分10
48秒前
欣阳1021完成签到,获得积分10
49秒前
今后应助SoaD采纳,获得10
49秒前
高分求助中
(应助此贴封号)【重要!!请各用户(尤其是新用户)详细阅读】【科研通的精品贴汇总】 10000
Kinesiophobia : a new view of chronic pain behavior 2000
Burger's Medicinal Chemistry, Drug Discovery and Development, Volumes 1 - 8, 8 Volume Set, 8th Edition 1800
Cronologia da história de Macau 1600
文献PREDICTION EQUATIONS FOR SHIPS' TURNING CIRCLES或期刊Transactions of the North East Coast Institution of Engineers and Shipbuilders第95卷 1000
BRITTLE FRACTURE IN WELDED SHIPS 1000
Lloyd's Register of Shipping's Approach to the Control of Incidents of Brittle Fracture in Ship Structures 1000
热门求助领域 (近24小时)
化学 材料科学 医学 生物 工程类 有机化学 纳米技术 计算机科学 化学工程 生物化学 物理 复合材料 内科学 催化作用 物理化学 光电子学 细胞生物学 基因 电极 遗传学
热门帖子
关注 科研通微信公众号,转发送积分 6150434
求助须知:如何正确求助?哪些是违规求助? 7979085
关于积分的说明 16574986
捐赠科研通 5262640
什么是DOI,文献DOI怎么找? 2808641
邀请新用户注册赠送积分活动 1788874
关于科研通互助平台的介绍 1656916