Reward-Independent Messaging for Decentralized Multi-Agent Reinforcement Learning
Yoshida, Naoto, Taniguchi, Tadahiro
–arXiv.org Artificial Intelligence
In multi-agent reinforcement learning (MARL), effective communication improves agent performance, particularly under partial observability. We propose MARL-CPC, a framework that enables communication among fully decentralized, independent agents without parameter sharing. MARL-CPC incorporates a message learning model based on collective predictive coding (CPC) from emergent communication research. Unlike conventional methods that treat messages as part of the action space and assume cooperation, MARL-CPC links messages to state inference, supporting communication in non-cooperative, reward-independent settings. We introduce two algorithms -Bandit-CPC and IPPO-CPC- and evaluate them in non-cooperative MARL tasks. Benchmarks show that both outperform standard message-as-action approaches, establishing effective communication even when messages offer no direct benefit to the sender. These results highlight MARL-CPC's potential for enabling coordination in complex, decentralized environments.
arXiv.org Artificial Intelligence
May-29-2025
- Country:
- Asia
- Japan > Honshū
- Chūbu > Nagano Prefecture
- Nagano (0.04)
- Kansai > Kyoto Prefecture
- Kyoto (0.05)
- Chūbu > Nagano Prefecture
- Middle East > Jordan (0.04)
- Japan > Honshū
- Asia
- Genre:
- Research Report (0.83)
- Industry:
- Health & Medicine (0.46)
- Technology: