Timezone: »
We develop theory and algorithms for average-reward on-policy Reinforcement Learning (RL). We first consider bounding the difference of the long-term average reward for two policies. We show that previous work based on the discounted return (Schulman et al. 2015, Achiam et al. 2017) results in a non-meaningful lower bound in the average reward setting. By addressing the average-reward criterion directly, we then derive a novel bound which depends on the average divergence between the policies and on Kemeny's constant. Based on this bound, we develop an iterative procedure which produces a sequence of monotonically improved policies for the average reward criterion. This iterative procedure can then be combined with classic Deep Reinforcement Learning (DRL) methods, resulting in practical DRL algorithms that target the long-run average reward criterion. In particular, we demonstrate that Average-Reward TRPO (ATRPO), which adapts the on-policy TRPO algorithm to the average-reward criterion, significantly outperforms TRPO in the most challenging MuJuCo environments.
Author Information
Yiming Zhang (New York University)
Keith Ross (New York University Shanghai)
Related Events (a corresponding poster, oral, or spotlight)
-
2021 Spotlight: On-Policy Deep Reinforcement Learning for the Average-Reward Criterion »
Wed. Jul 21st 12:30 -- 12:35 AM Room
More from the Same Authors
-
2020 Poster: Striving for Simplicity and Performance in Off-Policy DRL: Output Normalization and Non-Uniform Sampling »
Che Wang · Yanqiu Wu · Quan Vuong · Keith Ross