Goal-Oriented Multi-Agent Reinforcement Learning for Decentralized Agent Teams
Du, Hung, Nguyen, Hy, Thudumu, Srikanth, Vasa, Rajesh, Mouzakis, Kon
–arXiv.org Artificial Intelligence
Connected and autonomous vehicles across land, water, and air must often operate in dynamic, unpredictable environments with limited communication, no centralized control, and partial observability. These real-world constraints pose significant challenges for coordination, particularly when vehicles pursue individual objectives. To address this, we propose a decentralized Multi-Agent Reinforcement Learning (MARL) framework that enables vehicles, acting as agents, to communicate selectively based on local goals and observations. This goal-aware communication strategy allows agents to share only relevant information, enhancing collaboration while respecting visibility limitations. We validate our approach in complex multi-agent navigation tasks featuring obstacles and dynamic agent populations. Results show that our method significantly improves task success rates and reduces time-to-goal compared to non-cooperative baselines. Moreover, task performance remains stable as the number of agents increases, demonstrating scalability. These findings highlight the potential of decentralized, goal-driven MARL to support effective coordination in realistic multi-vehicle systems operating across diverse domains.
arXiv.org Artificial Intelligence
Nov-18-2025
- Country:
- Asia > Indonesia
- Bali (0.04)
- North America > United States
- Montana (0.04)
- Tennessee > Shelby County
- Germantown (0.04)
- Oceania > Australia (0.04)
- South America > Brazil
- São Paulo (0.04)
- Asia > Indonesia
- Genre:
- Research Report > New Finding (0.66)
- Industry:
- Leisure & Entertainment (0.46)