Skip to yearly menu bar Skip to main content


Poster
in
Workshop: Reinforcement Learning for Real Life

Objective Robustness in Deep Reinforcement Learning

Lauro Langosco di Langosco · Lee Sharkey


Abstract:

We study objective robustness failures, a type of out-of-distribution robustness failure in reinforcement learning (RL). Objective robustness failures occur when an RL agent retains its capabilities out-of-distribution yet pursues the wrong objective. This kind of failure presents different risks than the robustness problems usually considered in the literature, since it involves agents that leverage their capabilities to pursue the wrong objective rather than simply failing to do anything useful.We provide the first explicit empirical demonstrations of objective robustness failures and present a partial characterization of its causes.

Chat is not available.