Compositional Imitation Learning: Explaining and executing one task at a time

12/04/2018
by   Thomas Kipf, et al.
4

We introduce a framework for Compositional Imitation Learning and Execution (CompILE) of hierarchically-structured behavior. CompILE learns reusable, variable-length segments of behavior from demonstration data using a novel unsupervised, fully-differentiable sequence segmentation module. These learned behaviors can then be re-composed and executed to perform new tasks. At training time, CompILE auto-encodes observed behavior into a sequence of latent codes, each corresponding to a variable-length segment in the input sequence. Once trained, our model generalizes to sequences of longer length and from environment instances not seen during training. We evaluate our model in a challenging 2D multi-task environment and show that CompILE can find correct task boundaries and event encodings in an unsupervised manner without requiring annotated demonstration data. Latent codes and associated behavior policies discovered by CompILE can be used by a hierarchical agent, where the high-level policy selects actions in the latent code space, and the low-level, task-specific policies are simply the learned decoders. We found that our agent could learn given only sparse rewards, where agents without task-specific policies struggle.

READ FULL TEXT

page 7

page 12

page 15

page 16

research
03/21/2017

One-Shot Imitation Learning

Imitation learning has been commonly applied to solve different tasks in...
research
10/30/2019

Plan Arithmetic: Compositional Plan Vectors for Multi-Task Control

Autonomous agents situated in real-world environments must be able to ma...
research
03/28/2022

Modular Adaptive Policy Selection for Multi-Task Imitation Learning through Task Division

Deep imitation learning requires many expert demonstrations, which can b...
research
06/23/2021

Imitation Learning: Progress, Taxonomies and Opportunities

Imitation learning aims to extract knowledge from human experts' demonst...
research
06/17/2018

Learning Policy Representations in Multiagent Systems

Modeling agent behavior is central to understanding the emergence of com...
research
12/29/2019

Hierarchical Variational Imitation Learning of Control Programs

Autonomous agents can learn by imitating teacher demonstrations of the i...
research
12/14/2021

Learning to Guide and to Be Guided in the Architect-Builder Problem

We are interested in interactive agents that learn to coordinate, namely...

Please sign up or login with your details

Forgot password? Click here to reset