计算机科学
模式
模态(人机交互)
情态动词
推荐系统
图形
二部图
情报检索
人工智能
代表(政治)
机器学习
人机交互
多媒体
理论计算机科学
政治
政治学
社会学
化学
高分子化学
法学
社会科学
作者
Qifan Wang,Yinwei Wei,Jianhua Yin,Jianlong Wu,Xuemeng Song,Liqiang Nie
标识
DOI:10.1109/tmm.2021.3138298
摘要
One of the important factors affecting micro-video recommender systems is to model the multi-modal user preference on the micro-video. Despite the remarkable performance of prior arts, they are still limited by fusing the user preference derived from different modalities in a unified manner, ignoring the users tend to place different emphasis on different modalities. Furthermore, modality-missing is ubiquity and unavoidable in the micro-video recommendation, some modalities information of micro-videos are lacked in many cases, which negatively affects the multi-modal fusion operations. To overcome these disadvantages, we propose a novel framework for the micro-video recommendation, dubbed Dual Graph Neural Network (DualGNN), upon the user-microvideo bipartite and user co-occurrence graphs, which leverages the correlation between users to collaboratively mine the particular fusion pattern for each user. Specifically, we first introduce a single-modal representation learning module, which performs graph operations on the user-microvideo graph in each modality to capture single-modal user preferences on different modalities. And then, we devise a multi-modal representation learning module to explicitly model the user's attentions over different modalities and inductively learn the multi-modal user preference. Finally, we propose a prediction module to rank the potential micro-videos for users. Extensive experiments on two public datasets demonstrate the significant superiority of our DualGNN over state-of-the-arts methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI