Multi-Agent Common Knowledge Reinforcement Learning

Witt, Christian Schroeder de, Foerster, Jakob, Farquhar, Gregory, Torr, Philip, Boehmer, Wendelin, Whiteson, Shimon

Neural Information Processing Systems 

Cooperative multi-agent reinforcement learning often requires decentralised policies, which severely limit the agents' ability to coordinate their behaviour. In this paper, we show that common knowledge between agents allows for complex decentralised coordination. Common knowledge arises naturally in a large number of decentralised cooperative multi-agent tasks, for example, when agents can reconstruct parts of each others' observations. Since agents can independently agree on their common knowledge, they can execute complex coordinated policies that condition on this knowledge in a fully decentralised fashion. We propose multi-agent common knowledge reinforcement learning (MACKRL), a novel stochastic actor-critic algorithm that learns a hierarchical policy tree.