Research2026-05-14
Uncovering Symmetry Transfer in Large Language Models via Layer-Peeled Optimization
Source: Arxiv CS.AI
arXiv:2605.12756v1 Announce Type: cross Abstract: Large language models (LLMs) are pretrained by minimizing the cross-entropy loss for next-token prediction. In this paper, we study whether this optimization strategy can induce geometric structure in the learned model weights and context...
arxivpapers