AUPO -- Abstracted Until Proven Otherwise: A Reward Distribution Based Abstraction Algorithm
Reading time: 1 minute
...
📝 Original Info
- Title: AUPO – Abstracted Until Proven Otherwise: A Reward Distribution Based Abstraction Algorithm
- ArXiv ID: 2510.23214
- Date: 2025-10-27
- Authors: 정보 없음 (논문에 저자 정보가 제공되지 않았습니다.)
📝 Abstract
We introduce a novel, drop-in modification to Monte Carlo Tree Search's (MCTS) decision policy that we call AUPO. Comparisons based on a range of IPPC benchmark problems show that AUPO clearly outperforms MCTS. AUPO is an automatic action abstraction algorithm that solely relies on reward distribution statistics acquired during the MCTS. Thus, unlike other automatic abstraction algorithms, AUPO requires neither access to transition probabilities nor does AUPO require a directed acyclic search graph to build its abstraction, allowing AUPO to detect symmetric actions that state-of-the-art frameworks like ASAP struggle with when the resulting symmetric states are far apart in state space. Furthermore, as AUPO only affects the decision policy, it is not mutually exclusive with other abstraction techniques that only affect the tree search.💡 Deep Analysis
📄 Full Content
Reference
This content is AI-processed based on open access ArXiv data.