Research2026-05-12
SDG-MoE: Signed Debate Graph Mixture-of-Experts
Source: Arxiv CS.AI
arXiv:2605.08322v1 Announce Type: cross Abstract: Sparse MoE models achieve a good balance between capacity and compute by routing each token to a small subset of experts. However, in most MoE architectures, once a token is routed, the selected experts process it independently and their outputs are...
arxivpapers