Timezone: »

Recurrent Model-Free RL can be a Strong Baseline for Many POMDPs
Tianwei Ni · Benjamin Eysenbach · Ruslan Salakhutdinov

Wed Jul 20 03:30 PM -- 05:30 PM (PDT) @ Hall E #900

Many problems in RL, such as meta RL, robust RL, and generalization in RL, can be cast as POMDPs. In theory, simply augmenting model-free RL with memory-based architectures, such as recurrent neural networks, provides a general approach to solving all types of POMDPs. However, prior work has found that such recurrent model-free RL methods tend to perform worse than more specialized algorithms that are designed for specific types of POMDPs. This paper revisits this claim. We find that careful architecture and hyperparameter decisions can often yield a recurrent model-free implementation that performs on par with (and occasionally substantially better than) more sophisticated recent techniques. We compare to 19 environments from 5 prior specialized methods and find that our implementation achieves greater sample efficiency and asymptotic performance than these methods on 16 out of 19 environments. We also release a simple and efficient implementation of recurrent model-free RL for future work to use as a baseline for POMDPs.

Author Information

Tianwei Ni (Université de Montréal)
Benjamin Eysenbach (Carnegie Mellon University)
Ruslan Salakhutdinov (Carnegie Mellen University)

Related Events (a corresponding poster, oral, or spotlight)

More from the Same Authors