论文标题
训练深CTR模型的实用增量方法
A Practical Incremental Method to Train Deep CTR Models
论文作者
论文摘要
建议系统中的深度学习模型通常在批处理模式下进行训练,即在固定尺寸的训练数据窗口上进行迭代训练。这种深度学习模型的批处理模式训练遭受较低的训练效率,这可能会导致在未按时产生模型时性能下降。为了解决这个问题,提出了增量学习,最近受到了很多关注。增量学习在推荐系统中具有巨大的潜力,因为训练数据的两个连续窗口重叠了大部分卷。它的目的是在模型上次更新时仅使用时间戳新传入的样本来逐步更新模型,这比批处理模式培训要高得多。但是,大多数增量学习方法都集中在图像识别的研究领域,其中随着时间的流逝学习了新任务或类。在这项工作中,我们引入了一种实用的增量方法来训练深CTR模型,该模型由三个解耦模块(即数据,功能和模型模块)组成。我们的方法可以以更好的训练效率来实现与常规批处理模式培训相当的性能。我们对公共基准和私人数据集进行了广泛的实验,以证明我们提出的方法的有效性。
Deep learning models in recommender systems are usually trained in the batch mode, namely iteratively trained on a fixed-size window of training data. Such batch mode training of deep learning models suffers from low training efficiency, which may lead to performance degradation when the model is not produced on time. To tackle this issue, incremental learning is proposed and has received much attention recently. Incremental learning has great potential in recommender systems, as two consecutive window of training data overlap most of the volume. It aims to update the model incrementally with only the newly incoming samples from the timestamp when the model is updated last time, which is much more efficient than the batch mode training. However, most of the incremental learning methods focus on the research area of image recognition where new tasks or classes are learned over time. In this work, we introduce a practical incremental method to train deep CTR models, which consists of three decoupled modules (namely, data, feature and model module). Our method can achieve comparable performance to the conventional batch mode training with much better training efficiency. We conduct extensive experiments on a public benchmark and a private dataset to demonstrate the effectiveness of our proposed method.