Affordable Access

LEAF: Latent Exploration Along the Frontier

Authors
  • Bharadhwaj, Homanga
  • Garg, Animesh
  • Shkurti, Florian
Type
Preprint
Publication Date
Jun 18, 2020
Submission Date
May 21, 2020
Source
University of Michigan Library Repository
License
Yellow
External links

Abstract

Self-supervised goal proposal and reaching is a key component for exploration and efficient policy learning algorithms. Such a self-supervised approach without access to any oracle goal sampling distribution requires deep exploration and commitment so that long horizon plans can be efficiently discovered. In this paper, we propose an exploration framework, which learns a dynamics-aware manifold of reachable states. For a goal, our proposed method deterministically visits a state at the current frontier of reachable states (commitment/reaching) and then stochastically explores to reach the goal (exploration). This allocates exploration budget near the frontier of the reachable region instead of its interior. We target the challenging problem of policy learning from initial and goal states specified as images, and do not assume any access to the underlying ground-truth states of the robot and the environment. To keep track of reachable latent states, we propose a distance-conditioned reachability network that is trained to infer whether one state is reachable from another within the specified latent space distance. Given an initial state, we obtain a frontier of reachable states from that state. By incorporating a curriculum for sampling easier goals (closer to the start state) before more difficult goals, we demonstrate that the proposed self-supervised exploration algorithm, can achieve $20\%$ superior performance on average compared to existing baselines on a set of challenging robotic environments, including on a real robot manipulation task.

Report this publication

Statistics

Seen <100 times