Timezone: »

Preferential Temporal Difference Learning
Nishanth Anand · Doina Precup

Tue Jul 20 05:35 AM -- 05:40 AM (PDT) @ None

Temporal-Difference (TD) learning is a general and very useful tool for estimating the value function of a given policy, which in turn is required to find good policies. Generally speaking, TD learning updates states whenever they are visited. When the agent lands in a state, its value can be used to compute the TD-error, which is then propagated to other states. However, it may be interesting, when computing updates, to take into account other information than whether a state is visited or not. For example, some states might be more important than others (such as states which are frequently seen in a successful trajectory). Or, some states might have unreliable value estimates (for example, due to partial observability or lack of data), making their values less desirable as targets. We propose an approach to re-weighting states used in TD updates, both when they are the input and when they provide the target for the update. We prove that our approach converges with linear function approximation and illustrate its desirable empirical behaviour compared to other TD-style methods.

Author Information

Nishanth Anand (Mila / McGill University)
Doina Precup (McGill University / DeepMind)

Related Events (a corresponding poster, oral, or spotlight)

More from the Same Authors