对话
计算机科学
模式
超图
模态(人机交互)
图形
情绪分析
人工智能
话语
理论计算机科学
数学
社会科学
语言学
离散数学
哲学
社会学
作者
Jiaze Li,Hongyan Mei,Liyun Jia,Xing Zhang
出处
期刊:Electronics
[MDPI AG]
日期:2023-11-19
卷期号:12 (22): 4703-4703
被引量:4
标识
DOI:10.3390/electronics12224703
摘要
In recent years, sentiment analysis in conversation has garnered increasing attention due to its widespread applications in areas such as social media analytics, sentiment mining, and electronic healthcare. Existing research primarily focuses on sequence learning and graph-based approaches, yet they overlook the high-order interactions between different modalities and the long-term dependencies within each modality. To address these problems, this paper proposes a novel hypergraph-based method for multimodal emotion recognition in conversation (MER-HGraph). MER-HGraph extracts features from three modalities: acoustic, text, and visual. It treats each modality utterance in a conversation as a node and constructs intra-modal hypergraphs (Intra-HGraph) and inter-modal hypergraphs (Inter-HGraph) using hyperedges. The hypergraphs are then updated using hypergraph convolutional networks. Additionally, to mitigate noise in acoustic data and mitigate the impact of fixed time scales, we introduce a dynamic time window module to capture local-global information from acoustic signals. Extensive experiments on the IEMOCAP and MELD datasets demonstrate that MER-HGraph outperforms existing models in multimodal emotion recognition tasks, leveraging high-order information from multimodal data to enhance recognition capabilities.
科研通智能强力驱动
Strongly Powered by AbleSci AI