Abstract | ||
---|---|---|
Over the past decade, Monte Carlo Tree Search (MCTS) and specifically Upper Confidence Bound in Trees (UCT) have proven to be quite effective in large probabilistic planning domains. In this paper, we focus on how values are back-propagated in the MCTS tree, and apply complex return strategies from the Reinforcement Learning (RL) literature to MCTS, producing 4 new MCTS variants. We demonstrate that in some probabilistic planning benchmarks from the International Planning Competition (IPC), selecting a MCTS variant with a backup strategy different from Monte Carlo averaging can lead to substantially better results. We also propose a hypothesis for why different backup strategies lead to different performance in particular environments, and manipulate a carefully structured grid-world domain to provide empirical evidence supporting our hypothesis. |
Year | Venue | Field |
---|---|---|
2016 | ICML | Monte Carlo method,Monte Carlo tree search,Empirical evidence,Computer science,Artificial intelligence,Probabilistic logic,Machine learning,Backup,Reinforcement learning |
DocType | Citations | PageRank |
Conference | 4 | 0.40 |
References | Authors | |
16 | 4 |
Name | Order | Citations | PageRank |
---|---|---|---|
Piyush Khandelwal | 1 | 81 | 9.72 |
Elad Liebman | 2 | 21 | 5.69 |
S. Niekum | 3 | 165 | 23.73 |
Peter Stone | 4 | 6878 | 688.60 |