Learning to Collaborate in Markov Decision Processes
Goran Radanovic · Rati Devidze · David Parkes · Adish Singla

Tue Jun 11th 02:35 -- 02:40 PM @ Room 104
We consider a two-agent MDP framework where agents repeatedly solve a task in a collaborative setting. We study the problem of designing a learning algorithm for the first agent (A1) that facilitates a successful collaboration even in cases when the second agent (A2) is adapting its policy in an unknown way. The key challenge in our setting is that the presence of the second agent leads to non-stationarity and non-obliviousness of rewards and transitions for the first agent. We design novel online learning algorithms for agent A1 whose regret decays as $O(T^{1-\frac{3}{7} \cdot \alpha})$ with $T$ learning episodes provided that the magnitude of agent A2's policy changes between any two consecutive episodes are upper bounded by $O(T^{-\alpha})$. Here, the parameter $\alpha$ is assumed to be strictly greater than $0$, and we show that this assumption is necessary provided that the {\em learning parity with noise} problem is computationally hard. We show that sub-linear regret of agent A1 further implies near-optimality of the agents' joint return for MDPs that manifest the properties of a {\em smooth} game.

Author Information

Goran Radanovic (Harvard University)
Rati Devidze (Max Planck Institute for Software Systems)
David Parkes (Harvard University)
Adish Singla (Max Planck Institute (MPI-SWS))

Related Events (a corresponding poster, oral, or spotlight)

More from the Same Authors