DocumentCode
130218
Title
Converging to a player model in Monte-Carlo Tree Search
Author
Sarratt, Trevor ; Pynadath, David V. ; Jhala, Arnav
Author_Institution
Univ. of California at Santa Cruz, Santa Cruz, CA, USA
fYear
2014
fDate
26-29 Aug. 2014
Firstpage
1
Lastpage
7
Abstract
Player models allow search algorithms to account for differences in agent behavior according to player´s preferences and goals. However, it is often not until the first actions are taken that an agent can begin assessing which models are relevant to its current opponent. This paper investigates the integration of belief distributions over player models in the Monte-Carlo Tree Search (MCTS) algorithm. We describe a method of updating belief distributions through leveraging information sampled during the MCTS. We then characterize the effect of tuning parameters of the MCTS to convergence of belief distributions. Evaluation of this approach is done in comparison with value iteration for an iterated version of the prisoner´s dilemma problem. We show that for a sufficient quantity of iterations, our approach converges to the correct model faster than the same model under value iteration.
Keywords
Monte Carlo methods; belief maintenance; computer games; game theory; trees (mathematics); MCTS algorithm; Monte-Carlo tree search; agent behavior; belief distribution; player goals; player model convergence; player preference; prisoners dilemma problem; search algorithm; tuning parameter; value iteration; Backpropagation; Irrigation; Monte Carlo methods;
fLanguage
English
Publisher
ieee
Conference_Titel
Computational Intelligence and Games (CIG), 2014 IEEE Conference on
Conference_Location
Dortmund
Type
conf
DOI
10.1109/CIG.2014.6932881
Filename
6932881
Link To Document