虚拟现实
计算机科学
模式
卷积神经网络
领域(数学)
人机交互
人工智能
数学
社会科学
社会学
纯数学
作者
Yanfei Wang,Jingliang Wang,Lijun Wang,Zhengping Li,Ying Li
摘要
At present, emotion recognition has become a research hotspot in the field of pattern recognition. Considering the problems of incomplete information and strong interference in single‐modal emotion recognition, multimodal emotion recognition has been widely studied. Multimodal data includes, but is not limited to, emoji, text, and voice modality data. There are various ways to express emotion, among which expression, text and voice are the most direct and reliable emotional information carriers. Therefore, it is of great research and practical significance to comprehensively consider the emotion recognition research of expression, text and voice modalities, and to apply its research results to the field of virtual reality (referred to as VR). This paper analyzes the relevant situation of multimodal emotion recognition, extracts features of voice, text and expression, and then fuses them into multimodal for emotional analysis, and applies it to the VR field. The main work content is as follows: the relevant technologies of multimodal emotion recognition research in the field of VR are introduced, including deep learning related technologies, virtual reality technology, and multimodal fusion methods. In terms of deep learning, the focus is on convolutional neural networks and recurrent neural networks and their variants. In terms of virtual reality technology, the characteristics and applications of virtual reality are introduced. In terms of multimodal fusion, three commonly used fusion methods are introduced.
科研通智能强力驱动
Strongly Powered by AbleSci AI