Poster

Exploration in Approximate Hyper-State Space for Meta Reinforcement Learning

Luisa Zintgraf · Leo Feng · Cong Lu · Maximilian Igl · Kristian Hartikainen · Katja Hofmann · Shimon Whiteson

Virtual

Keywords: [ Algorithms ] [ Multitask, Transfer, and Meta Learning ]

[ Abstract ]
[ Slides
[ Paper ]
[ Visit Poster at Spot A6 in Virtual World ]
Thu 22 Jul 9 p.m. PDT — 11 p.m. PDT
 
Spotlight presentation: Multi-task Learning 1
Thu 22 Jul 5 p.m. PDT — 6 p.m. PDT

Abstract:

To rapidly learn a new task, it is often essential for agents to explore efficiently - especially when performance matters from the first timestep. One way to learn such behaviour is via meta-learning. Many existing methods however rely on dense rewards for meta-training, and can fail catastrophically if the rewards are sparse. Without a suitable reward signal, the need for exploration during meta-training is exacerbated. To address this, we propose HyperX, which uses novel reward bonuses for meta-training to explore in approximate hyper-state space (where hyper-states represent the environment state and the agent's task belief). We show empirically that HyperX meta-learns better task-exploration and adapts more successfully to new tasks than existing methods.

Chat is not available.