Oral Poster
SeeA*: Efficient Exploration-Enhanced A* Search by Selective Sampling
Dengwei Zhao · Shikui Tu · Lei Xu
West Ballroom A-D #6503
[
Abstract
]
[ Project Page ]
Oral
presentation:
Oral Session 2C: Reinforcement Learning
Wed 11 Dec 3:30 p.m. PST — 4:30 p.m. PST
Wed 11 Dec 4:30 p.m. PST
— 7:30 p.m. PST
Wed 11 Dec 3:30 p.m. PST — 4:30 p.m. PST
Abstract:
Monte-Carlo tree search (MCTS) and reinforcement learning contributed crucially to the success of AlphaGo and AlphaZero, and A$^*$ is a tree search algorithm among the most well-known ones in the classical AI literature. MCTS and A$^*$ both perform heuristic search and are mutually beneficial. Efforts have been made to the renaissance of A$^*$ from three possible aspects, two of which have been confirmed by studies in recent years, while the third is about the OPEN list that consists of open nodes of A$^*$ search, but still lacks deep investigation. This paper aims at the third, i.e., developing the Sampling-exploration enhanced A$^*$ (SeeA$^*$) search by constructing a dynamic subset of OPEN through a selective sampling process, such that the node with the best heuristic value in this subset instead of in the OPEN is expanded. Nodes with the best heuristic values in OPEN are most probably picked into this subset, but sometimes may not be included, which enables SeeA$^*$ to explore other promising branches. Three sampling techniques are presented for comparative investigations. Moreover, under the assumption about the distribution of prediction errors, we have theoretically shown the superior efficiency of SeeA$^*$ over A$^*$ search, particularly when the accuracy of the guiding heuristic function is insufficient. Experimental results on retrosynthetic planning in organic chemistry, logic synthesis in integrated circuit design, and the classical Sokoban game empirically demonstrate the efficiency of SeeA$^*$, in comparison with the state-of-the-art heuristic search algorithms.
Live content is unavailable. Log in and register to view live content