Research2026-04-30
TinyR1-32B-Preview: Boosting Accuracy with Branch-Merge Distillation
Source: Arxiv CS.AI
arXiv:2503.04872v3 Announce Type: replace-cross Abstract: The challenge of reducing the size of Large Language Models (LLMs) while maintaining their performance has gained significant attention. However, existing methods, such as model distillation and transfer learning, often fail to achieve high...
arxivpapers