学位论文详细信息
Solving planning problems with deep reinforcement learning and tree search
reinforcement learning;mcts;sokoban;a*;heuristic
Ge, Victor ; Lazebnik ; Svetlana
关键词: reinforcement learning;    mcts;    sokoban;    a*;    heuristic;   
Others  :  https://www.ideals.illinois.edu/bitstream/handle/2142/101086/GE-THESIS-2018.pdf?sequence=1&isAllowed=y
美国|英语
来源: The Illinois Digital Environment for Access to Learning and Scholarship
PDF
【 摘 要 】

Deep reinforcement learning methods are capable of learning complex heuristics starting with no prior knowledge, but struggle in environments where the learning signal is sparse. In contrast, planning methods can discover the optimal path to a goal in the absence of external rewards, but often require a hand-crafted heuristic function to be effective. In this thesis, we describe a model-based reinforcement learning method that bridges the middle ground between these two approaches. When evaluated on the complex domain of Sokoban, the model-based method was found to be more performant, stable and sample-efficient than a model-free baseline.

【 预 览 】
附件列表
Files Size Format View
Solving planning problems with deep reinforcement learning and tree search 656KB PDF download
  文献评价指标  
  下载次数:22次 浏览次数:15次