论文标题
量化机器人学习和概括的演示质量
Quantifying Demonstration Quality for Robot Learning and Generalization
论文作者
论文摘要
从示范中学习(LFD)试图通过使多样化的最终用户能够通过提供示范来教机器人执行任务来使机器人技术民主化。但是,大多数LFD技术都假定用户提供最佳的演示。在实际应用中,这种情况并非总是如此,因为用户可能会提供各种质量的演示,而这些质量可能会随着专业知识和其他因素而改变。演示质量在机器人学习和概括中起着至关重要的作用。因此,重要的是在将提供的演示的质量量化之前,然后将其用于机器人学习。在本文中,我们建议根据他们在学习任务中的表现来量化演示的质量。我们假设任务绩效可以指示相似任务上的概括性能。在用户研究中验证了所提出的方法(n = 27)。在不同的任务约束下,招募了具有不同机器人技术专业知识级别的用户来教PR2机器人一个通用任务(按下按钮)。他们在两个不同的日子里教了两次会议,以捕捉跨课程的教学行为。将任务绩效用于将提供的演示分类为高质量和低质量的集合。结果表明,所有参与者的任务绩效和概括性能之间的显着皮尔逊相关系数(r = 0.85,p <0.0001)。我们还发现,用户聚集在两个组中:提供了第一届会话的高质量演示,分配给快速适应器组的用户,以及在第一个会话中提供低质量演示的用户,然后通过练习进行了改进,并分配给了慢速适应器组。这些结果突出了量化演示质量的重要性,这可以指示用户对任务的适应水平。
Learning from Demonstration (LfD) seeks to democratize robotics by enabling diverse end-users to teach robots to perform a task by providing demonstrations. However, most LfD techniques assume users provide optimal demonstrations. This is not always the case in real applications where users are likely to provide demonstrations of varying quality, that may change with expertise and other factors. Demonstration quality plays a crucial role in robot learning and generalization. Hence, it is important to quantify the quality of the provided demonstrations before using them for robot learning. In this paper, we propose quantifying the quality of the demonstrations based on how well they perform in the learned task. We hypothesize that task performance can give an indication of the generalization performance on similar tasks. The proposed approach is validated in a user study (N = 27). Users with different robotics expertise levels were recruited to teach a PR2 robot a generic task (pressing a button) under different task constraints. They taught the robot in two sessions on two different days to capture their teaching behaviour across sessions. The task performance was utilized to classify the provided demonstrations into high-quality and low-quality sets. The results show a significant Pearson correlation coefficient (R = 0.85, p < 0.0001) between the task performance and generalization performance across all participants. We also found that users clustered into two groups: Users who provided high-quality demonstrations from the first session, assigned to the fast-adapters group, and users who provided low-quality demonstrations in the first session and then improved with practice, assigned to the slow-adapters group. These results highlight the importance of quantifying demonstration quality, which can be indicative of the adaptation level of the user to the task.