BeClaude
Policy2026-04-28

Meta-Aligner: Bidirectional Preference-Policy Optimization for Multi-Objective LLMs Alignment

Source: Arxiv CS.AI

arXiv:2604.24178v1 Announce Type: cross Abstract: Multi-Objective Alignment aims to align Large Language Models (LLMs) with diverse and often conflicting human values by optimizing multiple objectives simultaneously. Existing methods predominantly rely on static preference weight construction...

arxivpapers