K-Means Based TinyML Anomaly Detection and Distributed Model Reuse via the Distributed Internet of Learning (DIoL)

arXiv cs.LG / 3/31/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a lightweight K-Means-based anomaly detection pipeline for resource-constrained MCUs, including on-device feature extraction, clustering, and threshold estimation using real power measurements from a mini-fridge.
  • It introduces the Distributed Internet of Learning (DIoL) to export a trained model from one MCU as a portable, text-based representation for direct reuse on other devices without retraining.
  • A two-device prototype demonstrates the “Train Once, Share Everywhere” (TOSE) workflow, where one device trains and another performs inference using the shared model.
  • Experimental results indicate consistent anomaly detection behavior across devices, negligible parsing overhead, and identical inference runtimes compared with standalone (non-DIoL) operation.
  • The authors position the framework as a scalable approach for low-cost TinyML anomaly detection deployment across device fleets.

Abstract

This paper presents a lightweight K-Means anomaly detection model and a distributed model-sharing workflow designed for resource-constrained microcontrollers (MCUs). Using real power measurements from a mini-fridge appliance, the system performs on-device feature extraction, clustering, and threshold estimation to identify abnormal appliance behavior. To avoid retraining models on every device, we introduce the Distributed Internet of Learning (DIoL), which enables a model trained on one MCU to be exported as a portable, text-based representation and reused directly on other devices. A two-device prototype demonstrates the feasibility of the "Train Once, Share Everywhere" (TOSE) approach using a real-world appliance case study, where Device A trains the model and Device B performs inference without retraining. Experimental results show consistent anomaly detection behavior, negligible parsing overhead, and identical inference runtimes between standalone and DIoL-based operation. The proposed framework enables scalable, low-cost TinyML deployment across fleets of embedded devices.