眼电学
人工智能
计算机科学
脑电图
传感器融合
语音识别
情绪识别
融合
计算机视觉
模式识别(心理学)
心理学
神经科学
眼球运动
语言学
哲学
作者
Jialai Yin,Minchao Wu,Yan Yang,Ping Li,Fan Li,Wen Liang,Zhao Lv
标识
DOI:10.1109/tim.2024.3370813
摘要
Emotion recognition plays a vital role in building a harmonious society and emotional interaction. Recent research has demonstrated that multi-modal inter-channel correlations and insufficient emotion elicitation plague deep learning-based emotion identification techniques. To cope with these problems, we propose a transformer model of multi-modal and channel attention fusion (MCAF-Transformer). First, we employ an olfactory video approach to evoke emotional expression more fully and acquire EEG and EOG signal data. Second, the model makes full use of multi-modal channel information, time-domain and spatial-domain information of EEG and EOG signals, captures the correlation of different channels using channel attention, and improves the accuracy of emotion recognition by focusing on the global dependence on the temporal order using the transformer. We conducted extensive experiments on the olfactory video sentiment dataset, and the experimental results were correct at 94.63%. The results show that olfactory videos evoke emotion more adequately than pure videos and that the MCAF-Transformer model significantly outperforms other emotion recognition methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI