Learning World Graphs to Accelerate Hierarchical Reinforcement Learning

Subscribers:
284,000
Published on ● Video Link: https://www.youtube.com/watch?v=Qk4lJdp7ZAs



Duration: 18:39
3,111 views
99


The goal of hierarchical reinforcement learning is to divide a task into different levels of coarseness with the top-level agent planning only over a high-level view of the world and each subsequent layer having a more detailed view. This paper proposes to learn a set of important states as well as their connections to each other as a high-level abstraction.

https://arxiv.org/abs/1907.00664

Abstract:
In many real-world scenarios, an autonomous agent often encounters various tasks within a single complex environment. We propose to build a graph abstraction over the environment structure to accelerate the learning of these tasks. Here, nodes are important points of interest (pivotal states) and edges represent feasible traversals between them. Our approach has two stages. First, we jointly train a latent pivotal state model and a curiosity-driven goal-conditioned policy in a task-agnostic manner. Second, provided with the information from the world graph, a high-level Manager quickly finds solution to new tasks and expresses subgoals in reference to pivotal states to a low-level Worker. The Worker can then also leverage the graph to easily traverse to the pivotal states of interest, even across long distance, and explore non-locally. We perform a thorough ablation study to evaluate our approach on a suite of challenging maze tasks, demonstrating significant advantages from the proposed framework over baselines that lack world graph knowledge in terms of performance and efficiency.

Authors: Wenling Shang, Alex Trott, Stephan Zheng, Caiming Xiong, Richard Socher







Tags:
deep learning
reinforcement learning
deep reinforcement learning
world model
hierarchical reinforcement learning
planning
salesforce
research
machine learning
navigation
pivot states
ai
artificial intelligence