Timezone: »
The use of Reinforcement Learning (RL) agents in practical applications requires the consideration of suboptimal outcomes, depending on the familiarity of the agent with its environment. This is especially important in safety-critical environments, where errors can lead to high costs or damage. In distributional RL, the risk-sensitivity can be controlled via different distortion measures of the estimated return distribution. However, these distortion functions require an estimate of the risk level, which is difficult to obtain and depends on the current state. In this work, we demonstrate the suboptimality of a static risk level estimation and propose a method to dynamically select risk levels at each environment step. Our method ARA (Automatic Risk Adaptation) estimates the appropriate risk level in both known and unknown environments using a Random Network Distillation error. We show reduced failure rates by up to a factor of 7 and improved generalization performance by up to 14% compared to both risk-aware and risk-agnostic agents in several locomotion environments.
Author Information
Frederik Schubert (Leibniz University)
Theresa Eimer (Leibniz Universität Hannover)
Bodo Rosenhahn (Leibniz University Hannover)
Marius Lindauer (Leibniz University Hannover)
More from the Same Authors
-
2020 : Towards Self-Paced Context Evaluation for Contextual Reinforcement Learning »
Theresa Eimer -
2021 Workshop: 8th ICML Workshop on Automated Machine Learning (AutoML 2021) »
Gresa Shala · Frank Hutter · Joaquin Vanschoren · Marius Lindauer · Katharina Eggensperger · Colin White · Erin LeDell -
2021 Poster: Self-Paced Context Evaluation for Contextual Reinforcement Learning »
Theresa Eimer · André Biedenkapp · Frank Hutter · Marius Lindauer -
2021 Poster: TempoRL: Learning When to Act »
André Biedenkapp · Raghu Rajan · Frank Hutter · Marius Lindauer -
2021 Spotlight: TempoRL: Learning When to Act »
André Biedenkapp · Raghu Rajan · Frank Hutter · Marius Lindauer -
2021 Spotlight: Self-Paced Context Evaluation for Contextual Reinforcement Learning »
Theresa Eimer · André Biedenkapp · Frank Hutter · Marius Lindauer -
2020 Workshop: 7th ICML Workshop on Automated Machine Learning (AutoML 2020) »
Frank Hutter · Joaquin Vanschoren · Marius Lindauer · Charles Weill · Katharina Eggensperger · Matthias Feurer · Matthias Feurer -
2020 Poster: Lifted Disjoint Paths with Application in Multiple Object Tracking »
Andrea Hornakova · Roberto Henschel · Bodo Rosenhahn · Paul Swoboda