Back to News
Policy2026-04-17
TIP: Token Importance in On-Policy Distillation
Source: Arxiv CS.AI
arXiv:2604.14084v1 Announce Type: cross Abstract: On-policy knowledge distillation (OPD) trains a student on its own rollouts under token-level supervision from a teacher. Not all token positions matter equally, but existing views of token importance are incomplete. We ask a direct question: which...
arxivpapers