• DocumentCode
    130218
  • Title

    Converging to a player model in Monte-Carlo Tree Search

  • Author

    Sarratt, Trevor ; Pynadath, David V. ; Jhala, Arnav

  • Author_Institution
    Univ. of California at Santa Cruz, Santa Cruz, CA, USA
  • fYear
    2014
  • fDate
    26-29 Aug. 2014
  • Firstpage
    1
  • Lastpage
    7
  • Abstract
    Player models allow search algorithms to account for differences in agent behavior according to player´s preferences and goals. However, it is often not until the first actions are taken that an agent can begin assessing which models are relevant to its current opponent. This paper investigates the integration of belief distributions over player models in the Monte-Carlo Tree Search (MCTS) algorithm. We describe a method of updating belief distributions through leveraging information sampled during the MCTS. We then characterize the effect of tuning parameters of the MCTS to convergence of belief distributions. Evaluation of this approach is done in comparison with value iteration for an iterated version of the prisoner´s dilemma problem. We show that for a sufficient quantity of iterations, our approach converges to the correct model faster than the same model under value iteration.
  • Keywords
    Monte Carlo methods; belief maintenance; computer games; game theory; trees (mathematics); MCTS algorithm; Monte-Carlo tree search; agent behavior; belief distribution; player goals; player model convergence; player preference; prisoners dilemma problem; search algorithm; tuning parameter; value iteration; Backpropagation; Irrigation; Monte Carlo methods;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Computational Intelligence and Games (CIG), 2014 IEEE Conference on
  • Conference_Location
    Dortmund
  • Type

    conf

  • DOI
    10.1109/CIG.2014.6932881
  • Filename
    6932881