BeClaude
Policy2026-05-01

Accelerating Policy Synthesis in Large-Scale MDPs via Hierarchical Adaptive Refinement

Source: Arxiv CS.AI

arXiv:2506.17792v2 Announce Type: replace Abstract: Software-intensive systems, such as software product lines and robotics, utilise Markov decision processes (MDPs) to capture uncertainty and analyse sequential decision-making problems. Despite the usefulness of conventional policy synthesis...

arxivpapers