论文标题
如何重新训练推荐系统?顺序元学习方法
How to Retrain Recommender System? A Sequential Meta-Learning Method
论文作者
论文摘要
实用的推荐系统需要定期重新培训,以使用新的交互数据刷新模型。为了追求高模型保真度,通常希望在历史和新数据上重新审判该模型,因为它可以考虑长期和短期用户的偏好。但是,完整的模型再培训可能非常耗时和记忆成本,尤其是当历史数据的规模较大时。在这项工作中,我们研究了推荐系统的模型再培训机制,这是一个高实践价值的主题,但在研究界却很少探索。 我们的第一个信念是,对历史数据进行检验是不必要的,因为该模型以前已经对其进行了培训。然而,由于新数据的规模较小,并且包含有关长期用户偏好的信息较少,因此仅对新数据进行的正常培训可能很容易引起过度拟合和忘记问题。为了解决这一难题,我们提出了一种新的培训方法,旨在通过学习转移过去的培训经验来放弃历史数据。具体来说,我们设计了一个基于神经网络的转移组件,该组件将旧模型转换为一个针对未来建议的新模型。为了很好地了解转移组件,我们优化了“未来性能” - 即在下一个时期评估的建议精度。我们的顺序元学习(SML)方法提供了适用于任何可区分模型的通用训练范式。我们在矩阵分解上演示了SML,并在两个现实世界数据集上进行实验。经验结果表明,SML不仅实现了显着的加速,而且在建议准确性中胜过完整的模型再培训,从而验证了我们的建议的有效性。我们在以下网址发布我们的代码:https://github.com/zyang1580/sml。
Practical recommender systems need be periodically retrained to refresh the model with new interaction data. To pursue high model fidelity, it is usually desirable to retrain the model on both historical and new data, since it can account for both long-term and short-term user preference. However, a full model retraining could be very time-consuming and memory-costly, especially when the scale of historical data is large. In this work, we study the model retraining mechanism for recommender systems, a topic of high practical values but has been relatively little explored in the research community. Our first belief is that retraining the model on historical data is unnecessary, since the model has been trained on it before. Nevertheless, normal training on new data only may easily cause overfitting and forgetting issues, since the new data is of a smaller scale and contains fewer information on long-term user preference. To address this dilemma, we propose a new training method, aiming to abandon the historical data during retraining through learning to transfer the past training experience. Specifically, we design a neural network-based transfer component, which transforms the old model to a new model that is tailored for future recommendations. To learn the transfer component well, we optimize the "future performance" -- i.e., the recommendation accuracy evaluated in the next time period. Our Sequential Meta-Learning(SML) method offers a general training paradigm that is applicable to any differentiable model. We demonstrate SML on matrix factorization and conduct experiments on two real-world datasets. Empirical results show that SML not only achieves significant speed-up, but also outperforms the full model retraining in recommendation accuracy, validating the effectiveness of our proposals. We release our codes at: https://github.com/zyang1580/SML.