Timezone: »

Convergent Tree Backup and Retrace with Function Approximation
Ahmed Touati · Pierre-Luc Bacon · Doina Precup · Pascal Vincent

Thu Jul 12 09:15 AM -- 12:00 PM (PDT) @ Hall B #66

Off-policy learning is key to scaling up reinforcement learning as it allows to learn about a target policy from the experience generated by a different behavior policy. Unfortunately, it has been challenging to combine off-policy learning with function approximation and multi-step bootstrapping in a way that leads to both stable and efficient algorithms. In this work, we show that the Tree Backup and Retrace algorithms are unstable with linear function approximation, both in theory and in practice with specific examples. Based on our analysis, we then derive stable and efficient gradient-based algorithms using a quadratic convex-concave saddle-point formulation. By exploiting the problem structure proper to these algorithms, we are able to provide convergence guarantees and finite-sample bounds. The applicability of our new analysis also goes beyond Tree Backup and Retrace and allows us to provide new convergence rates for the GTD and GTD2 algorithms without having recourse to projections or Polyak averaging.

Author Information

Ahmed Touati (MILA / FAIR)
Pierre-Luc Bacon (McGill University)
Doina Precup (McGill University / DeepMind)
Pascal Vincent (U Montreal)

Related Events (a corresponding poster, oral, or spotlight)

More from the Same Authors