multi-agent training
Multi-Agent Training beyond Zero-Sum with Correlated Equilibrium Meta-Solvers
Marris, Luke, Muller, Paul, Lanctot, Marc, Tuyls, Karl, Grapael, Thore
Outside of normal form (NF) games, this problem setting Two-player, constant-sum games are well studied arises in multi-agent training when dealing with empirical in the literature, but there has been limited games (also called meta-games), where a game payoff progress outside of this setting. We propose Joint tensor is populated with expected outcomes between Policy-Space Response Oracles (JPSRO), an algorithm agents playing an extensive form (EF) game, for example for training agents in n-player, general-sum the StarCraft League (Vinyals et al., 2019) and Policy-Space extensive form games, which provably converges Response Oracles (PSRO) (Lanctot et al., 2017), a recent to an equilibrium. We further suggest correlated variant of which reached state-of-the-art results in Stratego equilibria (CE) as promising meta-solvers, and Barrage (McAleer et al., 2020).
- North America > United States > California > Monterey County > Monterey (0.04)
- North America > Puerto Rico > San Juan > San Juan (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- (2 more...)
How Alphabet's DeepMind used a 1999 video game to teach its AI teamwork
DeepMind, an Alphabet subsidiary, announced Tuesday its efforts to create an artificial intelligence (AI) that functions with human-like performance. Using Quake III Arena's Capture the Flag (CTF), a 3D first-person multiplayer video game, DeepMind taught AI how to work with and against humans to win the game. The rules of CTF haven't changed much since gym class. Two groups of individuals band together to steal the opponent's flag from their territory while also protecting their own. Teams can tag opponents that enter their territory, sending them back to their respective home base.