Research2026-04-28
Mixture of Heterogeneous Grouped Experts for Language Modeling
Source: Arxiv CS.AI
arXiv:2604.23108v1 Announce Type: cross Abstract: Large Language Models (LLMs) based on Mixture-of-Experts (MoE) are pivotal in industrial applications for their ability to scale performance efficiently. However, standard MoEs enforce uniform expert sizes,creating a rigidity that fails to align...
arxivpapers