推荐系统
计算机科学
强化学习
新颖性
推论
贝叶斯推理
利用
汤普森抽样
贪婪算法
任务(项目管理)
机器学习
贝叶斯概率
人工智能
算法
哲学
管理
经济
计算机安全
神学
作者
Wang Xin-xi,Yi Wang,David Hsu,Ye Wang
出处
期刊:Cornell University - arXiv
日期:2013-11-06
被引量:1
摘要
Current music recommender systems typically act in a greedy fashion by recommending songs with the highest user ratings. Greedy recommendation, however, is suboptimal over the long term: it does not actively gather information on user preferences and fails to recommend novel songs that are potentially interesting. A successful recommender system must balance the needs to explore user preferences and to exploit this information for recommendation. This paper presents a new approach to music recommendation by formulating this exploration-exploitation trade-off as a reinforcement learning task called the multi-armed bandit. To learn user preferences, it uses a Bayesian model, which accounts for both audio content and the novelty of recommendations. A piecewise-linear approximation to the model and a variational inference algorithm are employed to speed up Bayesian inference. One additional benefit of our approach is a single unified model for both music recommendation and playlist generation. Both simulation results and a user study indicate strong potential for the new approach.
科研通智能强力驱动
Strongly Powered by AbleSci AI