Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 80% Match Research Paper ML Engineers,Edge AI Developers,IoT Engineers,Researchers in Efficient ML,Mobile Developers 1 week ago

Memory Constrained Dynamic Subnetwork Update for Transfer Learning

generative-ai β€Ί diffusion
πŸ“„ Abstract

Abstract: On-device neural network training faces critical memory constraints that limit the adaptation of pre-trained models to downstream tasks. We present MeDyate, a theoretically-grounded framework for memory-constrained dynamic subnetwork adaptation. Our approach introduces two key innovations: LaRa (Layer Ranking), an improved layer importance metric that enables principled layer pre-selection, and a dynamic channel sampling strategy that exploits the temporal stability of channel importance distributions during fine-tuning. MeDyate dynamically resamples channels between epochs according to importance-weighted probabilities, ensuring comprehensive parameter space exploration while respecting strict memory budgets. Extensive evaluation across a large panel of tasks and architectures demonstrates that MeDyate achieves state-of-the-art performance under extreme memory constraints, consistently outperforming existing static and dynamic approaches while maintaining high computational efficiency. Our method represents a significant step towards enabling efficient on-device learning by demonstrating effective fine-tuning with memory budgets as low as a few hundred kB of RAM.
Authors (4)
AΓ«l QuΓ©lennec
Pavlo Mozharovskyi
Van-Tam Nguyen
Enzo Tartaglione
Submitted
October 23, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

Presents MeDyate, a theoretically-grounded framework for memory-constrained dynamic subnetwork adaptation on edge devices. It introduces Layer Ranking (LaRa) for principled layer pre-selection and a dynamic channel sampling strategy to respect strict memory budgets while enabling effective model adaptation.

Business Value

Enables powerful AI capabilities to run directly on resource-constrained devices (smartphones, IoT devices), reducing reliance on cloud infrastructure, improving privacy, and enabling real-time applications.