Policy2026-05-01
Accelerating Policy Synthesis in Large-Scale MDPs via Hierarchical Adaptive Refinement
Source: Arxiv CS.AI
arXiv:2506.17792v2 Announce Type: replace Abstract: Software-intensive systems, such as software product lines and robotics, utilise Markov decision processes (MDPs) to capture uncertainty and analyse sequential decision-making problems. Despite the usefulness of conventional policy synthesis...
arxivpapers