Timezone: »

Robust Subtask Learning for Compositional Generalization
Kishor Jothimurugan · Steve Hsu · Osbert Bastani · Rajeev Alur

Tue Jul 25 02:00 PM -- 04:30 PM (PDT) @ Exhibit Hall 1 #742

Compositional reinforcement learning is a promising approach for training policies to perform complex long-horizon tasks. Typically, a high-level task is decomposed into a sequence of subtasks and a separate policy is trained to perform each subtask. In this paper, we focus on the problem of training subtask policies in a way that they can be used to perform any task; here, a task is given by a sequence of subtasks. We aim to maximize the worst-case performance over all tasks as opposed to the average-case performance. We formulate the problem as a two agent zero-sum game in which the adversary picks the sequence of subtasks. We propose two RL algorithms to solve this game: one is an adaptation of existing multi-agent RL algorithms to our setting and the other is an asynchronous version which enables parallel training of subtask policies. We evaluate our approach on two multi-task environments with continuous states and actions and demonstrate that our algorithms outperform state-of-the-art baselines.

Author Information

Kishor Jothimurugan (Two Sigma)
Steve Hsu
Osbert Bastani (University of Pennsylvania)
Rajeev Alur (University of Pennsylvania)

More from the Same Authors