BeClaude
Research2026-05-12

Distributionally Robust Token Optimization in RLHF

Source: Arxiv CS.AI

arXiv:2604.08577v2 Announce Type: replace-cross Abstract: Large Language Models (LLMs) tend to respond correctly to prompts that align well with the data they were trained and fine-tuned on. Yet, small shifts in wording, format, or language can trigger surprisingly large failures, especially on...

arxivpapers