论文标题

公共政策的机器学习中可忽略的公平性 - 准确性权衡的经验观察

Empirical observation of negligible fairness-accuracy trade-offs in machine learning for public policy

论文作者

Rodolfa, Kit T., Lamba, Hemank, Ghani, Rayid

论文摘要

在政策和社会影响环境中使用机器学习的日益增长引起了人们对公平意义的关注,尤其是对于少数民族而言。这些问题引起了机器学习和人工智能研究人员的极大兴趣,他们开发了新的方法并建立了理论界限,以改善公平性,专注于源数据,正则化和模型培训或事后调整,以调整模型得分。但是,很少的工作研究了现实世界中的公平与准确性之间的实际权衡,以了解这些界限和方法如何转化为政策选择和对社会的影响。我们的实证研究通过调查减轻差异对准确性的影响来填补这一差距,重点是使用机器学习来为跨教育,心理健康,刑事司法和住房安全的资源受限计划中的福利分配提供福利分配。在这里,我们描述了应用工作,在这些工作中,我们发现公平准确的权衡在实践中可以忽略不计。在研究的每种环境中,明确专注于实现公平并使用我们提出的事后缓解方法,公平性得到了显着提高,而无需牺牲准确性。在研究的政策环境中,该观察结果是强大的,可用于干预群体的资源规模,时间和相对规模。这些经验结果挑战了一个普遍认为的假设,即减少差异需要接受明显的准确性下降或新颖,复杂的方法的发展,从而使这些应用中的差异更加实用。

Growing use of machine learning in policy and social impact settings have raised concerns for fairness implications, especially for racial minorities. These concerns have generated considerable interest among machine learning and artificial intelligence researchers, who have developed new methods and established theoretical bounds for improving fairness, focusing on the source data, regularization and model training, or post-hoc adjustments to model scores. However, little work has studied the practical trade-offs between fairness and accuracy in real-world settings to understand how these bounds and methods translate into policy choices and impact on society. Our empirical study fills this gap by investigating the impact of mitigating disparities on accuracy, focusing on the common context of using machine learning to inform benefit allocation in resource-constrained programs across education, mental health, criminal justice, and housing safety. Here we describe applied work in which we find fairness-accuracy trade-offs to be negligible in practice. In each setting studied, explicitly focusing on achieving equity and using our proposed post-hoc disparity mitigation methods, fairness was substantially improved without sacrificing accuracy. This observation was robust across policy contexts studied, scale of resources available for intervention, time, and relative size of the protected groups. These empirical results challenge a commonly held assumption that reducing disparities either requires accepting an appreciable drop in accuracy or the development of novel, complex methods, making reducing disparities in these applications more practical.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源