Distributed Value Decomposition Networks with Networked Agents
Varela, Guilherme S., Sardinha, Alberto, Melo, Francisco S.
–arXiv.org Artificial Intelligence
We investigate the problem of distributed training under partial observability, whereby cooperative multi-agent reinforcement learning agents (MARL) maximize the expected cumulative joint reward. We propose distributed value decomposition networks (DVDN) that generate a joint Q-function that factorizes into agent-wise Q-functions. Whereas the original value decomposition networks rely on centralized training, our approach is suitable for domains where centralized training is not possible and agents must learn by interacting with the physical environment in a decentralized manner while communicating with their peers. DVDN overcomes the need for centralized training by locally estimating the shared objective. We contribute with two innovative algorithms, DVDN and DVDN (GT), for the heterogeneous and homogeneous agents settings respectively. Empirically, both algorithms approximate the performance of value decomposition networks, in spite of the information loss during communication, as demonstrated in ten MARL tasks in three standard environments.
arXiv.org Artificial Intelligence
Feb-11-2025
- Country:
- Africa > Ethiopia
- Addis Ababa > Addis Ababa (0.04)
- Europe
- North America
- Canada > British Columbia
- Puerto Rico > San Juan
- San Juan (0.04)
- United States
- California > San Diego County
- San Diego (0.04)
- District of Columbia > Washington (0.04)
- Louisiana > Orleans Parish
- New Orleans (0.04)
- Massachusetts > Suffolk County
- Boston (0.04)
- New York > New York County
- New York City (0.04)
- California > San Diego County
- Oceania > New Zealand
- North Island > Auckland Region > Auckland (0.04)
- South America > Brazil
- Rio de Janeiro > Rio de Janeiro (0.04)
- Africa > Ethiopia
- Genre:
- Research Report (1.00)
- Industry:
- Leisure & Entertainment (0.93)