Research2026-05-12
Aligning Agents via Planning: A Benchmark for Trajectory-Level Reward Modeling
Source: Arxiv CS.AI
arXiv:2604.08178v2 Announce Type: replace Abstract: In classical Reinforcement Learning from Human Feedback (RLHF), Reward Models (RMs) serve as the fundamental signal provider for model alignment. As Large Language Models evolve into agentic systems capable of autonomous tool invocation and...
arxivpapersagentsbenchmark