论文标题

关于反事实解释器与推荐人之间的关系

On the Relationship Between Counterfactual Explainer and Recommender

论文作者

Liu, Gang, Zhang, Zhihan, Ning, Zheng, Jiang, Meng

论文摘要

推荐系统采用机器学习模型从历史数据中学习,以预测用户的偏好。深层神经网络(DNN)模型(例如神经协作过滤(NCF))越来越流行。但是,由于模型的复杂性和缺乏解释性,这些建议的有形性和可信赖性值得怀疑。为了启用解释性,诸如口音和国际汽联之类的最新技术正在寻找反事实解释,这些解释是用户的特定历史动作,删除的删除会导致建议结果改变。在这项工作中,我们为DNN和非DNN模型提供了一个一般框架,因此反事实解释器都属于它具有特定的组件选择。该框架首先估计了某个历史行动在删除后的影响,然后使用搜索算法来找到反事实解释的最小动作集。有了这个框架,我们能够研究解释者和推荐人之间的关系。我们经验研究了两个推荐模型(NCF和分解机)和两个数据集(Movielens和Yelp)。我们分析了推荐人的性能与解释器的质量之间的关系。我们观察到,通过标准评估指标,当建议更准确时,解释器的性能较差。这表明有良好的解释以纠正预测要比让他们进行错误的预测要难。社区需要更细粒度的评估指标,以衡量推荐系统的反事实解释的质量。

Recommender systems employ machine learning models to learn from historical data to predict the preferences of users. Deep neural network (DNN) models such as neural collaborative filtering (NCF) are increasingly popular. However, the tangibility and trustworthiness of the recommendations are questionable due to the complexity and lack of explainability of the models. To enable explainability, recent techniques such as ACCENT and FIA are looking for counterfactual explanations that are specific historical actions of a user, the removal of which leads to a change to the recommendation result. In this work, we present a general framework for both DNN and non-DNN models so that the counterfactual explainers all belong to it with specific choices of components. This framework first estimates the influence of a certain historical action after its removal and then uses search algorithms to find the minimal set of such actions for the counterfactual explanation. With this framework, we are able to investigate the relationship between the explainers and recommenders. We empirically study two recommender models (NCF and Factorization Machine) and two datasets (MovieLens and Yelp). We analyze the relationship between the performance of the recommender and the quality of the explainer. We observe that with standard evaluation metrics, the explainers deliver worse performance when the recommendations are more accurate. This indicates that having good explanations to correct predictions is harder than having them to wrong predictions. The community needs more fine-grained evaluation metrics to measure the quality of counterfactual explanations to recommender systems.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源