Timezone: »
Oral
Composing Value Functions in Reinforcement Learning
Benjamin van Niekerk · Steven James · Adam Earle · Benjamin Rosman
An important property for lifelong-learning agents is the ability to combine existing skills to solve new unseen tasks. In general, however, it is unclear how to compose existing skills in a principled manner. We show that optimal value function composition can be achieved in entropy-regularised reinforcement learning (RL), and then extend this result to the standard RL setting. Composition is demonstrated in a high-dimensional video game environment, where an agent with an existing library of skills is immediately able to solve new tasks without the need for further learning.
Author Information
Benjamin van Niekerk (University of the Witwatersrand)
Steve James (University of the Witwatersrand)
Adam Earle (University of the Witwatersrand)
Benjamin Rosman (Council for Scientific and Industrial Research)
Related Events (a corresponding poster, oral, or spotlight)
-
2019 Poster: Composing Value Functions in Reinforcement Learning »
Thu Jun 13th 01:30 -- 04:00 AM Room Pacific Ballroom
More from the Same Authors
-
2020 Poster: Learning Portable Representations for High-Level Planning »
Steven James · Benjamin Rosman · George Konidaris