Spotlight

Deep Coherent Exploration for Continuous Control

Yijie Zhang · Herke van Hoof

[ Abstract ] [ Livestream: Visit Reinforcement Learning 14 ] [ Paper ]
Wed 21 Jul 6:35 p.m. — 6:40 p.m. PDT

In policy search methods for reinforcement learning (RL), exploration is often performed by injecting noise either in action space at each step independently or in parameter space over each full trajectory. In prior work, it has been shown that with linear policies, a more balanced trade-off between these two exploration strategies is beneficial. However, that method did not scale to policies using deep neural networks. In this paper, we introduce deep coherent exploration, a general and scalable exploration framework for deep RL algorithms for continuous control, that generalizes step-based and trajectory-based exploration. This framework models the last layer parameters of the policy network as latent variables and uses a recursive inference step within the policy update to handle these latent variables in a scalable manner. We find that deep coherent exploration improves the speed and stability of learning of A2C, PPO, and SAC on several continuous control tasks.

Chat is not available.