Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: In multimodal learning, dominant modalities often overshadow others, limiting
generalization. We propose Modality-Aware Sharpness-Aware Minimization (M-SAM),
a model-agnostic framework that applies to many modalities and supports early
and late fusion scenarios. In every iteration, M-SAM in three steps optimizes
learning. \textbf{First, it identifies the dominant modality} based on
modalities' contribution in the accuracy using Shapley. \textbf{Second, it
decomposes the loss landscape}, or in another language, it modulates the loss
to prioritize the robustness of the model in favor of the dominant modality,
and \textbf{third, M-SAM updates the weights} by backpropagation of modulated
gradients. This ensures robust learning for the dominant modality while
enhancing contributions from others, allowing the model to explore and exploit
complementary features that strengthen overall performance. Extensive
experiments on four diverse datasets show that M-SAM outperforms the latest
state-of-the-art optimization and gradient manipulation methods and
significantly balances and improves multimodal learning.
Authors (4)
Hossein R. Nowdeh
Jie Ji
Xiaolong Ma
Fatemeh Afghah
Submitted
October 28, 2025
Key Contributions
Introduces Modality-Aware Sharpness-Aware Minimization (M-SAM), a model-agnostic framework that harmonizes multimodal learning by identifying dominant modalities and modulating gradients to improve robustness. It ensures better learning for underrepresented modalities, enhancing overall generalization.
Business Value
Enables the development of more reliable and versatile AI systems that can effectively process and integrate information from various sources (e.g., text, images, audio), leading to richer applications.