Combining Monte Carlo tree search and apprenticeship learning for capture the flag

Publisher:
IEEE
Publication Type:
Conference Proceeding
Citation:
2015 IEEE Conference on Computational Intelligence and Game, 2015, pp. 154 - 161
Issue Date:
2015-11-04
Full metadata record
Files in This Item:
Filename Description Size
07317914.pdfPublished version1.01 MB
Adobe PDF
© 2015 IEEE. In this paper we introduce a novel approach to agent control in competitive video games which combines Monte Carlo Tree Search (MCTS) and Apprenticeship Learning (AL). More specifically, an opponent model created through AL is used during the expansion phase of the Upper Confidence Bounds for Trees (UCT) variant of MCTS. We show how this approach can be applied to a game of Capture the Flag (CTF), an environment which is both non-deterministic and partially observable. The performance gain of a controller utilizing an opponent model learned via AL when compared to a controller using just UCT is shown both with win/loss ratios and True Skill rankings. Additionally, we build on previous findings by providing evidence of a bias towards a particular style of play in the AI Sandbox CTF environment. We believe that the approach highlighted here can be extended to a wider range of games other than just CTF.
Please use this identifier to cite or link to this item: