论文标题

选择,风险和奖励报告:为加强学习系统制定公共政策

Choices, Risks, and Reward Reports: Charting Public Policy for Reinforcement Learning Systems

论文作者

Gilbert, Thomas Krendl, Dean, Sarah, Zick, Tom, Lambert, Nathan

论文摘要

从长远来看,许多AI理论家认为增强学习(RL)是人工通用智能的最有前途的途径。这使RL从业人员可以设计从未存在过的系统,并且缺乏法律和政策中的先前文件。公共机构可以干预以前过于不透明而无法考虑的复杂动态,而长期以来的政策野心最终将成为可解决的问题。在此白皮书中,我们说明了这种潜力以及如何在能源基础设施,社交媒体推荐系统和运输领域进行技术制定。除了这些前所未有的干预措施外,还有新的风险形式,加剧了标准机器学习工具已经产生的危害。我们相应地提出了由RL设计选择引起的新风险类型,属于四个类别:范围范围,定义奖励,修剪信息和培训多个代理。政策制定者不允许RL系统单方面重塑人类领域,而是需要新的机制来实现理性,可预见性和互操作性,以使这些系统构成风险。我们认为,这些选择的标准可以从反托拉斯,侵权和行政法中的新兴子场中得出。然后,法院,联邦和州机构以及非政府组织将有可能在RL规范和评估中发挥更积极的作用。建立在Mitchell等人提出的“模型卡”和“数据表”框架上。和Gebru等人,我们认为需要对AI系统的奖励报告。奖励报告是针对划定设计选择的建议RL部署的生存文件。

In the long term, reinforcement learning (RL) is considered by many AI theorists to be the most promising path to artificial general intelligence. This places RL practitioners in a position to design systems that have never existed before and lack prior documentation in law and policy. Public agencies could intervene on complex dynamics that were previously too opaque to deliberate about, and long-held policy ambitions would finally be made tractable. In this whitepaper we illustrate this potential and how it might be technically enacted in the domains of energy infrastructure, social media recommender systems, and transportation. Alongside these unprecedented interventions come new forms of risk that exacerbate the harms already generated by standard machine learning tools. We correspondingly present a new typology of risks arising from RL design choices, falling under four categories: scoping the horizon, defining rewards, pruning information, and training multiple agents. Rather than allowing RL systems to unilaterally reshape human domains, policymakers need new mechanisms for the rule of reason, foreseeability, and interoperability that match the risks these systems pose. We argue that criteria for these choices may be drawn from emerging subfields within antitrust, tort, and administrative law. It will then be possible for courts, federal and state agencies, and non-governmental organizations to play more active roles in RL specification and evaluation. Building on the "model cards" and "datasheets" frameworks proposed by Mitchell et al. and Gebru et al., we argue the need for Reward Reports for AI systems. Reward Reports are living documents for proposed RL deployments that demarcate design choices.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源