Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: In recent years, artificial intelligence has significantly advanced medical
image segmentation. Nonetheless, challenges remain, including efficient 3D
medical image processing across diverse modalities and handling data
variability. In this work, we introduce Hierarchical Soft Mixture-of-Experts
(HoME), a two-level token-routing layer for efficient long-context modeling,
specifically designed for 3D medical image segmentation. Built on the Mamba
Selective State Space Model (SSM) backbone, HoME enhances sequential modeling
through adaptive expert routing. In the first level, a Soft Mixture-of-Experts
(SMoE) layer partitions input sequences into local groups, routing tokens to
specialized per-group experts for localized feature extraction. The second
level aggregates these outputs through a global SMoE layer, enabling
cross-group information fusion and global context refinement. This hierarchical
design, combining local expert routing with global expert refinement, enhances
generalizability and segmentation performance, surpassing state-of-the-art
results across datasets from the three most widely used 3D medical imaging
modalities and varying data qualities. The code is publicly available at
https://github.com/gmum/MambaHoME.
Authors (5)
Szymon Płotka
Gizem Mert
Maciej Chrabaszcz
Ewa Szczurek
Arkadiusz Sitek
Key Contributions
This paper proposes HoME, a novel hierarchical Soft Mixture-of-Experts layer built upon the Mamba SSM backbone, specifically for efficient 3D medical image segmentation. HoME enables long-context modeling through a two-level token-routing mechanism that combines localized feature extraction with global context refinement, addressing challenges in processing diverse 3D medical data.
Business Value
Improves the accuracy and efficiency of medical image analysis, leading to better diagnostic tools and potentially faster drug discovery and treatment planning.