Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 95% Match Research Paper LLM researchers,Machine learning engineers,AI practitioners 1 week ago

S'MoRE: Structural Mixture of Residual Experts for Parameter-Efficient LLM Fine-tuning

large-language-models › model-architecture
📄 Abstract

Abstract: Fine-tuning pre-trained large language models (LLMs) presents a dual challenge of balancing parameter efficiency and model capacity. Existing methods like low-rank adaptations (LoRA) are efficient but lack flexibility, while Mixture-of-Experts (MoE) enhance model capacity at the cost of more & under-utilized parameters. To address these limitations, we propose Structural Mixture of Residual Experts (S'MoRE), a novel framework that seamlessly integrates the efficiency of LoRA with the flexibility of MoE. Conceptually, S'MoRE employs hierarchical low-rank decomposition of expert weights, yielding residuals of varying orders interconnected in a multi-layer structure. By routing input tokens through sub-trees of residuals, S'MoRE emulates the capacity of numerous experts by instantiating and assembling just a few low-rank matrices. We craft the inter-layer propagation of S'MoRE's residuals as a special type of Graph Neural Network (GNN), and prove that under similar parameter budget, S'MoRE improves structural flexibility of traditional MoE (or Mixture-of-LoRA) by exponential order. Comprehensive theoretical analysis and empirical results demonstrate that S'MoRE achieves superior fine-tuning performance, offering a transformative approach for efficient LLM adaptation. Our implementation is available at: https://github.com/ZimpleX/SMoRE-LLM.
Authors (10)
Hanqing Zeng
Yinglong Xia
Zhuokai Zhao
Chuan Jiang
Qiang Zhang
Jiayi Liu
+4 more
Submitted
April 8, 2025
arXiv Category
cs.CL
arXiv PDF

Key Contributions

S'MoRE is a novel framework that combines the parameter efficiency of LoRA with the capacity of MoE by employing hierarchical low-rank decomposition of expert weights. It routes tokens through sub-trees of residuals, emulating many experts with few low-rank matrices, and formulates inter-layer propagation using GNNs.

Business Value

Enables organizations to fine-tune and deploy powerful LLMs more cost-effectively, democratizing access to advanced AI capabilities and allowing for rapid adaptation to specific business needs.