计算机科学
水准点(测量)
模棱两可
光学(聚焦)
模式
人工智能
过程(计算)
一致性(知识库)
多模态
机器学习
自然语言处理
人机交互
万维网
社会学
地理
程序设计语言
大地测量学
物理
光学
操作系统
社会科学
作者
Zheng Lian,Licai Sun,Mingyu Xu,Haiyang Sun,Ke Xu,Zhuofan Wen,Shun Chen,Bin Liu,Jianhua Tao
出处
期刊:Cornell University - arXiv
日期:2023-01-01
被引量:3
标识
DOI:10.48550/arxiv.2306.15401
摘要
Multimodal emotion recognition is an active research topic in artificial intelligence. Its primary objective is to integrate multi-modalities (such as acoustic, visual, and lexical clues) to identify human emotional states. Current works generally assume accurate emotion labels for benchmark datasets and focus on developing more effective architectures. But due to the inherent subjectivity of emotions, existing datasets often lack high annotation consistency, resulting in potentially inaccurate labels. Consequently, models built on these datasets may struggle to meet the demands of practical applications. To address this issue, it is crucial to enhance the reliability of emotion annotations. In this paper, we propose a novel task called ``\textbf{Explainable Multimodal Emotion Reasoning (EMER)}''. In contrast to previous works that primarily focus on predicting emotions, EMER takes a step further by providing explanations for these predictions. The prediction is considered correct as long as the reasoning process behind the predicted emotion is plausible. This paper presents our initial efforts on EMER, where we introduce a benchmark dataset, establish baseline models, and define evaluation metrics. Meanwhile, we observe the necessity of integrating multi-faceted capabilities to deal with EMER. Therefore, we propose the first multimodal large language model (LLM) in affective computing, called \textbf{AffectGPT}. We aim to tackle the long-standing challenge of label ambiguity and chart a path toward more reliable techniques. Furthermore, EMER offers an opportunity to evaluate the audio-video-text understanding capabilities of recent multimodal LLM. To facilitate further research, we make the code and data available at: https://github.com/zeroQiaoba/AffectGPT.
科研通智能强力驱动
Strongly Powered by AbleSci AI