Timezone: »
Spotlight
Sample-Efficient Reinforcement Learning with loglog(T) Switching Cost
Dan Qiao · Ming Yin · Ming Min · Yu-Xiang Wang
Thu Jul 21 08:30 AM -- 08:35 AM (PDT) @ Ballroom 3 & 4
We study the problem of reinforcement learning (RL) with low (policy) switching cost — a problem well-motivated by real-life RL applications in which deployments of new policies are costly and the number of policy updates must be low. In this paper, we propose a new algorithm based on stage-wise exploration and adaptive policy elimination that achieves a regret of $\widetilde{O}(\sqrt{H^4S^2AT})$ while requiring a switching cost of $O(HSA \log\log T)$. This is an exponential improvement over the best-known switching cost $O(H^2SA\log T)$ among existing methods with $\widetilde{O}(\mathrm{poly}(H,S,A)\sqrt{T})$ regret. In the above, $S,A$ denotes the number of states and actions in an $H$-horizon episodic Markov Decision Process model with unknown transitions, and $T$ is the number of steps. As a byproduct of our new techniques, we also derive a reward-free exploration algorithm with a switching cost of $O(HSA)$. Furthermore, we prove a pair of information-theoretical lower bounds which say that (1) Any no-regret algorithm must have a switching cost of $\Omega(HSA)$; (2) Any $\widetilde{O}(\sqrt{T})$ regret algorithm must incur a switching cost of $\Omega(HSA\log\log T)$. Both our algorithms are thus optimal in their switching costs.
Author Information
Dan Qiao (UCSB)
Ming Yin (UC Santa Barbara)
Ming Min (University of California, Santa Barbara)
Yu-Xiang Wang (UC Santa Barbara)
Related Events (a corresponding poster, oral, or spotlight)
-
2022 Poster: Sample-Efficient Reinforcement Learning with loglog(T) Switching Cost »
Thu. Jul 21st through Fri the 22nd Room Hall E #1117
More from the Same Authors
-
2021 : Optimal Uniform OPE and Model-based Offline Reinforcement Learning in Time-Homogeneous, Reward-Free and Task-Agnostic Settings »
Ming Yin · Yu-Xiang Wang -
2021 : Near-Optimal Offline Reinforcement Learning via Double Variance Reduction »
Ming Yin · Yu Bai · Yu-Xiang Wang -
2022 : Optimal Dynamic Regret in LQR Control »
Dheeraj Baby · Yu-Xiang Wang -
2021 Poster: Signatured Deep Fictitious Play for Mean Field Games with Common Noise »
Ming Min · Ruimeng Hu -
2021 Spotlight: Signatured Deep Fictitious Play for Mean Field Games with Common Noise »
Ming Min · Ruimeng Hu