MOMA-AC: A preference-driven actor-critic framework for continuous multi-objective multi-agent reinforcement learning
Callaghan, Adam, Mason, Karl, Mannion, Patrick
–arXiv.org Artificial Intelligence
This paper addresses a critical gap in Multi-Objective Multi-Agent Reinforcement Learning (MOMARL) by introducing the first dedicated inner-loop actor-critic framework for continuous state and action spaces: Multi-Objective Multi-Agent Actor-Critic (MOMA-AC). Building on single-objective, single-agent algorithms, we instantiate this framework with Twin Delayed Deep Deterministic Policy Gradient (TD3) and Deep Deterministic Policy Gradient (DDPG), yielding MOMA-TD3 and MOMA-DDPG. The framework combines a multi-headed actor network, a centralised critic, and an objective preference-conditioning architecture, enabling a single neural network to encode the Pareto front of optimal trade-off policies for all agents across conflicting objectives in a continuous MOMARL setting. We also outline a natural test suite for continuous MOMARL by combining a pre-existing multi-agent single-objective physics simulator with its multi-objective single-agent counterpart. Evaluating cooperative locomotion tasks in this suite, we show that our framework achieves statistically significant improvements in expected utility and hypervolume relative to outer-loop and independent training baselines, while demonstrating stable scalability as the number of agents increases. These results establish our framework as a foundational step towards robust, scalable multi-objective policy learning in continuous multi-agent domains.
arXiv.org Artificial Intelligence
Nov-25-2025
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe > Ireland
- Connaught > County Galway > Galway (0.04)
- South America > Chile
- Asia > Middle East
- Genre:
- Research Report > Experimental Study (0.46)
- Industry:
- Energy (0.93)
- Transportation > Ground
- Road (0.46)