A Unified Game-Theoretic Approach to Multiagent Reinforcement Learning
–Neural Information Processing Systems
To achieve general intelligence, agents must learn how to interact with others in a shared environment: this is the challenge of multiagent reinforcement learning (MARL). The simplest form is independent reinforcement learning (InRL), where each agent treats its experience as part of its (non-stationary) environment. In this paper, we first observe that policies learned using InRL can overfit to the other agents' policies during training, failing to sufficiently generalize during execution. We introduce a new metric, joint-policy correlation, to quantify this effect. We describe an algorithm for general MARL, based on approximate best responses to mixtures of policies generated using deep reinforcement learning, and empirical game-theoretic analysis to compute meta-strategies for policy selection.
Neural Information Processing Systems
Oct-3-2024, 03:09:13 GMT
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe > United Kingdom
- England > Cambridgeshire > Cambridge (0.04)
- North America
- Canada
- United States
- California
- Los Angeles County
- Long Beach (0.04)
- Los Angeles (0.14)
- San Francisco County > San Francisco (0.14)
- Los Angeles County
- New Jersey > Middlesex County
- New Brunswick (0.04)
- New York (0.04)
- Texas > Travis County
- Austin (0.04)
- California
- Oceania > Australia
- New South Wales > Sydney (0.04)
- Asia > Middle East
- Genre:
- Research Report (0.46)
- Industry:
- Leisure & Entertainment > Games > Computer Games (0.93)
- Technology: