论文标题
人类代理合作的温暖和能力
Warmth and competence in human-agent cooperation
论文作者
论文摘要
与人类的互动和合作是人工智能(AI)研究的总体愿望。最近的研究表明,接受深入增强学习训练的AI代理能够与人类合作。这些研究主要通过“客观”指标(例如任务绩效)评估人类兼容性,从而掩盖了不同代理人Garner的信任水平和主观偏好水平的潜在变化。为了更好地了解人类代理合作中主观偏好的因素,我们培训了两人社会困境硬币中的深入强化学习者。我们招募$ n = 501 $参与者进行人类合作研究,并衡量他们对遇到的代理商的印象。参与者对温暖和能力的看法预测了他们对不同代理的偏好,超出了客观绩效指标。从社会科学和生物学研究中汲取灵感,我们随后实施了一个新的``合作选择''框架来引发偏好:在与经纪人一起播放一集之后,询问参与者是否想与同一代理商一起玩下一集或独自播放。与既定的偏好一样,社会感知可以更好地预测参与者所揭示的偏好,而不是客观的绩效。鉴于这些结果,我们建议人类代理人的互动研究人员通常将社会感知和主观偏好的测量纳入他们的研究中。
Interaction and cooperation with humans are overarching aspirations of artificial intelligence (AI) research. Recent studies demonstrate that AI agents trained with deep reinforcement learning are capable of collaborating with humans. These studies primarily evaluate human compatibility through "objective" metrics such as task performance, obscuring potential variation in the levels of trust and subjective preference that different agents garner. To better understand the factors shaping subjective preferences in human-agent cooperation, we train deep reinforcement learning agents in Coins, a two-player social dilemma. We recruit $N = 501$ participants for a human-agent cooperation study and measure their impressions of the agents they encounter. Participants' perceptions of warmth and competence predict their stated preferences for different agents, above and beyond objective performance metrics. Drawing inspiration from social science and biology research, we subsequently implement a new ``partner choice'' framework to elicit revealed preferences: after playing an episode with an agent, participants are asked whether they would like to play the next episode with the same agent or to play alone. As with stated preferences, social perception better predicts participants' revealed preferences than does objective performance. Given these results, we recommend human-agent interaction researchers routinely incorporate the measurement of social perception and subjective preferences into their studies.