Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 60% Match Research Paper Embedded systems engineers,IoT developers,ML researchers focusing on edge AI,CPS researchers 1 week ago

RockNet: Distributed Learning on Ultra-Low-Power Devices

graph-neural-networks โ€บ graph-learning
๐Ÿ“„ Abstract

Abstract: As Machine Learning (ML) becomes integral to Cyber-Physical Systems (CPS), there is growing interest in shifting training from traditional cloud-based to on-device processing (TinyML), for example, due to privacy and latency concerns. However, CPS often comprise ultra-low-power microcontrollers, whose limited compute resources make training challenging. This paper presents RockNet, a new TinyML method tailored for ultra-low-power hardware that achieves state-of-the-art accuracy in timeseries classification, such as fault or malware detection, without requiring offline pretraining. By leveraging that CPS consist of multiple devices, we design a distributed learning method that integrates ML and wireless communication. RockNet leverages all devices for distributed training of specialized compute efficient classifiers that need minimal communication overhead for parallelization. Combined with tailored and efficient wireless multi-hop communication protocols, our approach overcomes the communication bottleneck that often occurs in distributed learning. Hardware experiments on a testbed with 20 ultra-low-power devices demonstrate RockNet's effectiveness. It successfully learns timeseries classification tasks from scratch, surpassing the accuracy of the latest approach for neural network microcontroller training by up to 2x. RockNet's distributed ML architecture reduces memory, latency and energy consumption per device by up to 90 % when scaling from one central device to 20 devices. Our results show that a tight integration of distributed ML, distributed computing, and communication enables, for the first time, training on ultra-low-power hardware with state-of-the-art accuracy.
Authors (4)
Alexander Grรคfe
Fabian Mager
Marco Zimmerling
Sebastian Trimpe
Submitted
October 15, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

RockNet is a novel TinyML method enabling distributed learning on ultra-low-power microcontrollers for CPS. It integrates ML and wireless communication, allowing multiple devices to collaboratively train specialized, compute-efficient classifiers without offline pretraining, achieving state-of-the-art accuracy with minimal communication overhead.

Business Value

Enables intelligent capabilities directly on low-cost, low-power edge devices, reducing reliance on cloud infrastructure, enhancing privacy, and enabling real-time decision-making for applications like predictive maintenance and security monitoring.