论文标题

学习动态的抽象表示,用于样品有效的增强学习

Learning Dynamic Abstract Representations for Sample-Efficient Reinforcement Learning

论文作者

Dadvar, Mehdi, Nayyar, Rashmeet Kaur, Srivastava, Siddharth

论文摘要

在许多现实世界中,学习代理需要同时学习问题的抽象和解决方案。但是,大多数此类抽象需要手工设计和完善,以解决不同问题和应用领域。本文提出了一种新颖的自上而下的方法,用于在进行强化学习时构建状态抽象。从状态变量和模拟器开始,它提出了一种与域无关的方法,用于动态地计算基于Q值在抽象状态中的Q值的分散,因为代理人继续进行和学习。对多个领域和问题的广泛经验评估表明,这种方法会自动学习对问题进行精细调整的抽象,产生强大的样本效率,并导致RL代理显着优于现有方法。

In many real-world problems, the learning agent needs to learn a problem's abstractions and solution simultaneously. However, most such abstractions need to be designed and refined by hand for different problems and domains of application. This paper presents a novel top-down approach for constructing state abstractions while carrying out reinforcement learning. Starting with state variables and a simulator, it presents a novel domain-independent approach for dynamically computing an abstraction based on the dispersion of Q-values in abstract states as the agent continues acting and learning. Extensive empirical evaluation on multiple domains and problems shows that this approach automatically learns abstractions that are finely-tuned to the problem, yield powerful sample efficiency, and result in the RL agent significantly outperforming existing approaches.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源