Timezone: »

Challenges and Opportunities in Offline Reinforcement Learning from Visual Observations
Cong Lu · Philip Ball · Tim G. J Rudner · Jack Parker-Holder · Michael A Osborne · Yee-Whye Teh
Event URL: https://openreview.net/forum?id=lyIe2Ki58zZ »

Offline reinforcement learning has shown great promise in leveraging large pre-collected datasets for policy learning, allowing agents to forgo often-expensive online data collection. However, to date, offline reinforcement learning from visual observations with continuous action spaces has been relatively under-explored, and there is a lack of understanding of where the remaining challenges lie. In this paper, we seek to establish simple baselines for continuous control in the visual domain. We show that simple modifications to two state-of-the-art vision-based online reinforcement learning algorithms, DreamerV2 and DrQ-v2, suffice to outperform prior work and establish a competitive baseline. We rigorously evaluate these algorithms on both existing offline datasets and a new testbed for offline reinforcement learning from visual observations that better represents the data distributions present in real-world offline RL problems, and open-source our code and data to facilitate progress in this important domain. Finally, we present and analyze several key desiderata unique to offline RL from visual observations, including visual distractions and visually identifiable changes in dynamics.

Author Information

Cong Lu (University of Oxford)
Philip Ball (University of Oxford)
Tim G. J Rudner (University of Oxford)
Jack Parker-Holder (University of Oxford)
Michael A Osborne (U Oxford)
Yee-Whye Teh (Oxford and DeepMind)

More from the Same Authors