BeClaude
Research2026-04-22

Distillation Traps and Guards: A Calibration Knob for LLM Distillability

Source: Arxiv CS.AI

arXiv:2604.18963v1 Announce Type: cross Abstract: Knowledge distillation (KD) transfers capabilities from large language models (LLMs) to smaller students, yet it can fail unpredictably and also underpins model leakage risks. Our analysis revealed several distillation traps: tail noise, off-policy...

arxivpapers