论文标题
使用语言模型的几个次数次目标计划
Few-shot Subgoal Planning with Language Models
论文作者
论文摘要
预先训练的大语言模型在许多语言理解基准的许多语言中都取得了成功。这项工作探讨了这些模型在现实环境中预测可行计划的能力。考虑到文本指令,我们表明在预训练的语言模型中编码的语言先验使我们能够推断出细粒度的亚目标序列。与最近对亚目标监督做出有力假设的方法相反,我们的实验表明,语言模型可以从少数训练序列中推断出详细的亚目标序列而无需进行任何微调。我们进一步提出了一个简单的策略,以基于环境的互动和反馈来重新排列语言模型预测。与先前的方法相比,与先前的基准方法相比,我们的方法与预先训练的导航和视觉推理组件相结合,在Alfred基准测试中表现出竞争性表现。
Pre-trained large language models have shown successful progress in many language understanding benchmarks. This work explores the capability of these models to predict actionable plans in real-world environments. Given a text instruction, we show that language priors encoded in pre-trained language models allow us to infer fine-grained subgoal sequences. In contrast to recent methods which make strong assumptions about subgoal supervision, our experiments show that language models can infer detailed subgoal sequences from few training sequences without any fine-tuning. We further propose a simple strategy to re-rank language model predictions based on interaction and feedback from the environment. Combined with pre-trained navigation and visual reasoning components, our approach demonstrates competitive performance on subgoal prediction and task completion in the ALFRED benchmark compared to prior methods that assume more subgoal supervision.