论文标题
使用用户不断发展的偏好分解的顺序推荐
Sequential Recommendation with User Evolving Preference Decomposition
论文作者
论文摘要
建模用户顺序行为最近引起了推荐域中越来越多的关注。现有方法主要假设同一序列相干偏好。但是,用户个性是动荡的且容易改变的,并且用户行为可能存在多种混合偏好。为了解决这个问题,在本文中,我们通过分解和建模用户独立的偏好提出了一种新颖的顺序推荐模型。为了实现这一目标,我们重点介绍了三个实用的挑战,考虑到用户行为的不一致,不断发展且不平衡的性质,这很少被先前的工作注意到。为了克服统一框架中的这些挑战,我们引入了一个增强学习模块,以模拟用户偏好的演变。更具体地说,该动作旨在根据以前的项目分解方式以及连续行为之间的时间间隔,将每个项目分配为子序列或创建新项目。奖励与最终的学习目标丧失有关,旨在产生可以更好地适合培训数据的子序列。我们基于不同领域的六个现实世界数据集进行了广泛的实验。与最先进的方法相比,经验研究表明,我们的模型平均可以分别对精度,召回,NDCG和MRR的指标分别提高约8.21%,10.08%,10.32%和9.82%。
Modeling user sequential behaviors has recently attracted increasing attention in the recommendation domain. Existing methods mostly assume coherent preference in the same sequence. However, user personalities are volatile and easily changed, and there can be multiple mixed preferences underlying user behaviors. To solve this problem, in this paper, we propose a novel sequential recommender model via decomposing and modeling user independent preferences. To achieve this goal, we highlight three practical challenges considering the inconsistent, evolving and uneven nature of the user behavior, which are seldom noticed by the previous work. For overcoming these challenges in a unified framework, we introduce a reinforcement learning module to simulate the evolution of user preference. More specifically, the action aims to allocate each item into a sub-sequence or create a new one according to how the previous items are decomposed as well as the time interval between successive behaviors. The reward is associated with the final loss of the learning objective, aiming to generate sub-sequences which can better fit the training data. We conduct extensive experiments based on six real-world datasets across different domains. Compared with the state-of-the-art methods, empirical studies manifest that our model can on average improve the performance by about 8.21%, 10.08%, 10.32%, and 9.82% on the metrics of Precision, Recall, NDCG and MRR, respectively.