Towards Energy Impact on AI-Powered 6G IoT Networks: Centralized vs. Decentralized

arXiv cs.AI / 4/22/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The study analyzes energy consumption trade-offs for machine-learning workloads in 6G IoT networks under centralized versus decentralized (distributed) architectures, focusing on training and data transmission costs.
  • It reports results from a real testbed deployed in Germany’s railway infrastructure, using sensor-driven ML for predictive maintenance.
  • Comparative evaluation shows that distributed learning can achieve competitive predictive accuracy of about 90% while cutting total electricity consumption by as much as 70%.
  • The findings suggest distributed ML is a practical approach to improving energy efficiency in real-world IoT deployments, particularly by reducing transmission-related energy overhead.

Abstract

The emergence of sixth-generation (6G) technologies has introduced new challenges and opportunities for machine learning (ML) applications in Internet of Things (IoT) networks, particularly concerning energy efficiency. As model training and data transmission contribute significantly to energy consumption, optimizing these processes has become critical for sustainable system design. This study first conduct analysis on the energy consumption model for both centralized and decentralized architecture and then presents a testbed deployed within the German railway infrastructure, leveraging sensor data for ML-based predictive maintenance. A comparative analysis of distributed versus Centralized Learning (CL) architectures reveals that distributed models maintain competitive predictive accuracy (~90%) while reducing overall electricity consumption by up to 70%. These findings underscore the potential of distributed ML to improve energy efficiency in real-world IoT deployments, particularly by mitigating transmission-related energy costs.