Timezone: »

The Virtues of Laziness in Model-based RL: A Unified Objective and Algorithms
Anirudh Vemula · Yuda Song · Aarti Singh · J. Bagnell · Sanjiban Choudhury

Tue Jul 25 02:00 PM -- 04:30 PM (PDT) @ Exhibit Hall 1 #428
Event URL: https://github.com/vvanirudh/LAMPS-MBRL »

We propose a novel approach to addressing two fundamental challenges in Model-based Reinforcement Learning (MBRL): the computational expense of repeatedly finding a good policy in the learned model, and the objective mismatch between model fitting and policy computation. Our "lazy" method leverages a novel unified objective, Performance Difference via Advantage in Model, to capture the performance difference between the learned policy and expert policy under the true dynamics. This objective demonstrates that optimizing the expected policy advantage in the learned model under an exploration distribution is sufficient for policy computation, resulting in a significant boost in computational efficiency compared to traditional planning methods. Additionally, the unified objective uses a value moment matching term for model fitting, which is aligned with the model's usage during policy computation. We present two no-regret algorithms to optimize the proposed objective, and demonstrate their statistical and computational gains compared to existing MBRL methods through simulated benchmarks.

Author Information

Anirudh Vemula (Aurora Innovation)
Yuda Song (Carnegie Mellon University)
Aarti Singh (Carnegie Mellon University)
J. Bagnell (Aurora Innovation)
Sanjiban Choudhury (Cornell University)

More from the Same Authors