论文标题

使用视觉排列的动作测序

Action sequencing using visual permutations

论文作者

Burke, Michael, Subr, Kartic, Ramamoorthy, Subramanian

论文摘要

人类可以轻松地推理完成任务所需的高级动作顺序,但是很难将这种能力灌输在相对较少的示例中训练的机器人中。这项工作考虑了以单个参考视觉状态为条件的神经动作测序的任务。该任务极具挑战性,因为它不仅受到大型动作集产生的重要组合复杂性的影响,而且还需要一个模型,该模型可以执行某种形式的符号接地,并将高维输入数据映射到动作,同时推理动作关系。本文采用排列的观点,并认为对动作测序的作用有益于推理排列和订购概念的能力。经验分析表明,在受约束的动作测序任务中,接受潜在排列训练的神经模型优于标准神经体系结构。结果还表明,使用视觉排列的动作测序是初始化和加快传统计划技术并成功扩展到比以前考虑的模型更大的动作集大小的有效机制。

Humans can easily reason about the sequence of high level actions needed to complete tasks, but it is particularly difficult to instil this ability in robots trained from relatively few examples. This work considers the task of neural action sequencing conditioned on a single reference visual state. This task is extremely challenging as it is not only subject to the significant combinatorial complexity that arises from large action sets, but also requires a model that can perform some form of symbol grounding, mapping high dimensional input data to actions, while reasoning about action relationships. This paper takes a permutation perspective and argues that action sequencing benefits from the ability to reason about both permutations and ordering concepts. Empirical analysis shows that neural models trained with latent permutations outperform standard neural architectures in constrained action sequencing tasks. Results also show that action sequencing using visual permutations is an effective mechanism to initialise and speed up traditional planning techniques and successfully scales to far greater action set sizes than models considered previously.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源