Timezone: »
We present a framework for autonomously learning a portable representation that describes a collection of low-level continuous environments. We show that these abstract representations can be learned in a task-independent egocentric space specific to the agent that, when grounded with problem-specific information, are provably sufficient for planning. We demonstrate transfer in two different domains, where an agent learns a portable, task-independent symbolic vocabulary, as well as operators expressed in that vocabulary, and then learns to instantiate those operators on a per-task basis. This reduces the number of samples required to learn a representation of a new task.
Author Information
Steven James (University of the Witwatersrand)
Benjamin Rosman (University of the Witwatersrand, South Africa)
Benjamin Rosman received a Ph.D. degree in Informatics from the University of Edinburgh in 2014. Previously, he obtained an M.Sc. in Artificial Intelligence from the University of Edinburgh, a Bachelor of Science (Honours) in Computer Science from the University of the Witwatersrand, South Africa, and a Bachelor of Science (Honours) in Computational and Applied Mathematics, also from the University of the Witwatersrand. He is an Associate Professor in the School of Computer Science and Applied Mathematics at the University of the Witwatersrand. He is the Chair of the IEEE South African joint chapter of Control Systems, and Robotics and Automation. Prof. Rosman’s research interests focus primarily on learning and decision making in autonomous systems, in particular studying how learning can be accelerated through abstracting and generalising knowledge gained from solving previous problems. He additionally works in the area of skill and behaviour learning for robots.
George Konidaris (Brown)
More from the Same Authors
-
2023 : Guided Policy Search for Parameterized Skills using Adverbs »
Benjamin Spiegel · George Konidaris -
2023 Poster: Flipping Coins to Estimate Pseudocounts for Exploration in Reinforcement Learning »
Sam Lobel · Akhil Bagaria · George Konidaris -
2023 Oral: Flipping Coins to Estimate Pseudocounts for Exploration in Reinforcement Learning »
Sam Lobel · Akhil Bagaria · George Konidaris -
2023 Poster: Meta-learning Parameterized Skills »
Haotian Fu · Shangqun Yu · Saket Tiwari · Michael L. Littman · George Konidaris -
2023 Poster: RLang: A Declarative Language for Describing Partial World Knowledge to Reinforcement Learning Agents »
Rafael A Rodriguez-Sanchez · Benjamin Spiegel · Jennifer Wang · Roma Patel · Stefanie Tellex · George Konidaris -
2022 : Just-in-Time Sparsity: Learning Dynamic Sparsity Schedules »
· Chiratidzo Matowe · Arnu Pretorius · Benjamin Rosman · Sara Hooker -
2021 : RL + Robotics Panel »
George Konidaris · Jan Peters · Martin Riedmiller · Angela Schoellig · Rose Yu · Rupam Mahmood -
2021 Poster: Skill Discovery for Exploration and Planning using Deep Skill Graphs »
Akhil Bagaria · Jason Senthil · George Konidaris -
2021 Oral: Skill Discovery for Exploration and Planning using Deep Skill Graphs »
Akhil Bagaria · Jason Senthil · George Konidaris -
2020 : Learning Object-Centric Representations for High-Level Planning in Minecraft »
Steven James -
2019 : Benjamin Rosman: Exploiting Structure For Accelerating Reinforcement Learning »
Benjamin Rosman -
2019 Poster: Finding Options that Minimize Planning Time »
Yuu Jinnai · David Abel · David Hershkowitz · Michael L. Littman · George Konidaris -
2019 Oral: Finding Options that Minimize Planning Time »
Yuu Jinnai · David Abel · David Hershkowitz · Michael L. Littman · George Konidaris -
2019 Poster: Composing Value Functions in Reinforcement Learning »
Benjamin van Niekerk · Steven James · Adam Earle · Benjamin Rosman -
2019 Oral: Composing Value Functions in Reinforcement Learning »
Benjamin van Niekerk · Steven James · Adam Earle · Benjamin Rosman -
2019 Poster: Discovering Options for Exploration by Minimizing Cover Time »
Yuu Jinnai · Jee Won Park · David Abel · George Konidaris -
2019 Oral: Discovering Options for Exploration by Minimizing Cover Time »
Yuu Jinnai · Jee Won Park · David Abel · George Konidaris -
2018 Poster: Policy and Value Transfer in Lifelong Reinforcement Learning »
David Abel · Yuu Jinnai · Sophie Guo · George Konidaris · Michael L. Littman -
2018 Oral: Policy and Value Transfer in Lifelong Reinforcement Learning »
David Abel · Yuu Jinnai · Sophie Guo · George Konidaris · Michael L. Littman -
2017 Poster: Hierarchy Through Composition with Multitask LMDPs »
Andrew Saxe · Adam Earle · Benjamin Rosman -
2017 Talk: Hierarchy Through Composition with Multitask LMDPs »
Andrew Saxe · Adam Earle · Benjamin Rosman