Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cl 90% Match Research Paper ML researchers,AI architects,Deep learning engineers 1 week ago

Unified Sparse Mixture of Experts

large-language-models › model-architecture
📄 Abstract

Abstract: Sparse Mixture of Experts (SMoEs) models scale the capacity of models while maintaining constant computational overhead. Early designs typically relied on a fixed value of $k$, where $k$ represents either the number of experts selected per token or the number of tokens assigned per expert. However, these approaches encounter three key limitations: they may fail to route to important experts or tokens, may assign irrelevant ones, and often suffer from representation collapse among experts. This paper reexamines SMoEs through the lens of \textit{Linear Programming}, and proposes a Unified Sparse Mixture of Experts (USMoE) framework that addresses these limitations. Specifically, our approach introduces a unified mechanism that integrates information from both the expert and token dimensions, and a unified scoring function that linearly combines similarity scores between experts and tokens. We provide both theoretical justification and empirical evidence demonstrating USMoE's effectiveness in overcoming the limitations of traditional routing methods. Through comprehensive evaluations on both clean and corrupted settings for large language models and vision tasks, under both training-free and training scenarios, USMoE achieves up to a 10\% performance improvement over standard approaches or reduces inference costs by up to 14\%, while maintaining competitive accuracy.
Authors (3)
Giang Do
Hung Le
Truyen Tran
Submitted
March 29, 2025
arXiv Category
cs.CL
arXiv PDF

Key Contributions

Proposes a Unified Sparse Mixture of Experts (USMoE) framework that re-examines SMoEs through Linear Programming. USMoE introduces a unified mechanism integrating expert and token dimensions and a unified scoring function to address limitations of fixed k-SMoEs, such as poor routing and representation collapse.

Business Value

Enables the development of larger, more capable models with controlled computational costs, leading to more powerful AI applications in NLP and beyond.