计算机科学
判别式
人工智能
微博
社会化媒体
卷积神经网络
机器学习
概率逻辑
噪音(视频)
利用
模式
深度学习
可扩展性
情绪分析
光学(聚焦)
模式识别(心理学)
图像(数学)
光学
物理
万维网
数据库
社会学
计算机安全
社会科学
作者
Fuhai Chen,Rongrong Ji,Jinsong Su,Donglin Cao,Yue Gao
标识
DOI:10.1109/tmm.2017.2757769
摘要
Predicting sentiments of multimodal microblogs composed of text, image, and emoticon have attracted ever-increasing research focus recently. The key challenge lies in the difficulty of collecting a sufficient amount of training labels to train a discriminative model for multimodal prediction. One potential solution is to exploit the labels collected from social media users, which is, however, restricted by the negative effect of label noise. Besides, we have quantitatively found that sentiments in different modalities may be independent, which disables the usage of previous multimodal sentiment analysis schemes in our problem. In this paper, we introduce a weakly supervised multimodal deep learning (WS-MDL) scheme toward robust and scalable sentiment prediction. WS-MDL learns convolutional neural networks iteratively and selectively from “weak” emoticon labels, which are cheaply available and noise containing. In particular, to filter out the label noise and to capture the modality dependency, a probabilistic graphical model is introduced to simultaneously learn discriminative multimodal descriptors and infer the confidence of label noise. Extensive evaluations are conducted in a million-scale, real-world microblog sentiment dataset crawled from Sina Weibo. We have validated the merits of the proposed scheme by quantitatively showing its superior performance over several state-of-the-art and alternative approaches.
科研通智能强力驱动
Strongly Powered by AbleSci AI