计算机科学
电影
利用
杠杆(统计)
情态动词
情报检索
模式
图形
代表(政治)
人工智能
推荐系统
万维网
人机交互
协同过滤
多媒体
理论计算机科学
政治学
社会学
政治
化学
高分子化学
法学
计算机安全
社会科学
作者
Yinwei Wei,Xiang Wang,Liqiang Nie,Xiangnan He,Richang Hong,Tat‐Seng Chua
出处
期刊:ACM Multimedia
日期:2019-10-15
卷期号:: 1437-1445
被引量:553
标识
DOI:10.1145/3343031.3351034
摘要
Personalized recommendation plays a central role in many online content sharing platforms. To provide quality micro-video recommendation service, it is of crucial importance to consider the interactions between users and items (i.e. micro-videos) as well as the item contents from various modalities (e.g. visual, acoustic, and textual). Existing works on multimedia recommendation largely exploit multi-modal contents to enrich item representations, while less effort is made to leverage information interchange between users and items to enhance user representations and further capture user's fine-grained preferences on different modalities. In this paper, we propose to exploit user-item interactions to guide the representation learning in each modality, and further personalized micro-video recommendation. We design a Multi-modal Graph Convolution Network (MMGCN) framework built upon the message-passing idea of graph neural networks, which can yield modal-specific representations of users and micro-videos to better capture user preferences. Specifically, we construct a user-item bipartite graph in each modality, and enrich the representation of each node with the topological structure and features of its neighbors. Through extensive experiments on three publicly available datasets, Tiktok, Kwai, and MovieLens, we demonstrate that our proposed model is able to significantly outperform state-of-the-art multi-modal recommendation methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI