Timezone: »
We consider the problem of the limited-bandwidth communication for multi-agent reinforcement learning, where agents cooperate with the assistance of a communication protocol and a scheduler. The protocol and scheduler jointly determine which agent is communicating what message and to whom. Under the limited bandwidth constraint, a communication protocol is required to generate informative messages. Meanwhile, an unnecessary communication connection should not be established because it occupies limited resources in vain. In this paper, we develop an Informative Multi-Agent Communication (IMAC) method to learn efficient communication protocols as well as scheduling. First, from the perspective of communication theory, we prove that the limited bandwidth constraint requires low-entropy messages throughout the transmission. Then inspired by the information bottleneck principle, we learn a valuable and compact communication protocol and a weight-based scheduler. To demonstrate the efficiency of our method, we conduct extensive experiments in various cooperative and competitive multi-agent tasks with different numbers of agents and different bandwidths. We show that IMAC converges faster and leads to efficient communication among agents under the limited bandwidth as compared to many baseline methods.
Author Information
Rundong Wang (Nanyang Technological University)
Xu He (Nanyang Technological University)
Runsheng Yu (Nanyang Technological University)
Wei Qiu (Nanyang Technological University)
Bo An (Nanyang Technological University)
Zinovi Rabinovich (Nanyang Technological University)
More from the Same Authors
-
2021 : Contingency-Aware Influence Maximization: A Reinforcement Learning Approach »
Haipeng Chen · Wei Qiu · Han-Ching Ou · Bo An · Milind Tambe -
2022 Poster: Mitigating Neural Network Overconfidence with Logit Normalization »
Hongxin Wei · RENCHUNZI XIE · Hao Cheng · LEI FENG · Bo An · Yixuan Li -
2022 Poster: Learning Pseudometric-based Action Representations for Offline Reinforcement Learning »
Pengjie Gu · Mengchen Zhao · Chen Chen · Dong Li · Jianye Hao · Bo An -
2022 Spotlight: Learning Pseudometric-based Action Representations for Offline Reinforcement Learning »
Pengjie Gu · Mengchen Zhao · Chen Chen · Dong Li · Jianye Hao · Bo An -
2022 Spotlight: Mitigating Neural Network Overconfidence with Logit Normalization »
Hongxin Wei · RENCHUNZI XIE · Hao Cheng · LEI FENG · Bo An · Yixuan Li -
2022 Poster: Open-Sampling: Exploring Out-of-Distribution data for Re-balancing Long-tailed datasets »
Hongxin Wei · Lue Tao · RENCHUNZI XIE · LEI FENG · Bo An -
2022 Spotlight: Open-Sampling: Exploring Out-of-Distribution data for Re-balancing Long-tailed datasets »
Hongxin Wei · Lue Tao · RENCHUNZI XIE · LEI FENG · Bo An -
2021 Poster: Pointwise Binary Classification with Pairwise Confidence Comparisons »
Lei Feng · Senlin Shu · Nan Lu · Bo Han · Miao Xu · Gang Niu · Bo An · Masashi Sugiyama -
2021 Poster: Learning from Similarity-Confidence Data »
Yuzhou Cao · Lei Feng · Yitian Xu · Bo An · Gang Niu · Masashi Sugiyama -
2021 Spotlight: Learning from Similarity-Confidence Data »
Yuzhou Cao · Lei Feng · Yitian Xu · Bo An · Gang Niu · Masashi Sugiyama -
2021 Spotlight: Pointwise Binary Classification with Pairwise Confidence Comparisons »
Lei Feng · Senlin Shu · Nan Lu · Bo Han · Miao Xu · Gang Niu · Bo An · Masashi Sugiyama -
2020 Poster: Learning with Multiple Complementary Labels »
LEI FENG · Takuo Kaneko · Bo Han · Gang Niu · Bo An · Masashi Sugiyama