The Combinatorial Multi-Armed Bandit Problem and Its Application to Real-Time Strategy Games
Ontanon, Santiago (Drexel University)
Game tree search in games with large branching factors is a notoriously hard problem. In this paper, we address this problem with a new sampling strategy for Monte Carlo Tree Search (MCTS) algorithms, called "Naive Sampling", based on a variant of the Multi-armed Bandit problem called the "Combinatorial Multi-armed Bandit" (CMAB) problem. We present a new MCTS algorithm based on Naive Sampling called NaiveMCTS, and evaluate it in the context of real-time strategy (RTS) games. Our results show that as the branching factor grows, NaiveMCTS performs significantly better than other algorithms.
Nov-10-2013
- Genre:
- Research Report > New Finding (0.53)
- Industry:
- Leisure & Entertainment > Games > Computer Games (0.40)
- Technology: