AI Navigate

Balancing Performance and Fairness in Explainable AI for Anomaly Detection in Distributed Power Plants Monitoring

arXiv cs.LG / 3/20/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies anomaly detection in distributed power plant monitoring, focusing on extreme class imbalance, interpretability, and fairness across regional clusters, using a dataset of diesel generator operations in Cameroon.
  • It presents a supervised ML framework that combines ensemble models (LightGBM, XGBoost, Random Forest, CatBoost, GBDT, AdaBoost) with baseline models (SVM, K-NN, MLP, Logistic Regression) and advanced resampling (SMOTE with Tomek Links and ENN) to address imbalance.
  • It uses SHAP for interpretability and the Disparate Impact Ratio (DIR) to quantify fairness, with Maximum Mean Discrepancy (MMD) to assess domain shifts between regions; results show ensemble models outperform baselines and LightGBM achieves F1 around 0.99 with DIR near 0.95.
  • It discusses deployment for real-time monitoring via containerized services, enabling low-latency predictions and actionable, interpretable outputs for operators.

Abstract

Reliable anomaly detection in distributed power plant monitoring systems is essential for ensuring operational continuity and reducing maintenance costs, particularly in regions where telecom operators heavily rely on diesel generators. However, this task is challenged by extreme class imbalance, lack of interpretability, and potential fairness issues across regional clusters. In this work, we propose a supervised ML framework that integrates ensemble methods (LightGBM, XGBoost, Random Forest, CatBoost, GBDT, AdaBoost) and baseline models (Support Vector Machine, K-Nearrest Neighbors, Multilayer Perceptrons, and Logistic Regression) with advanced resampling techniques (SMOTE with Tomek Links and ENN) to address imbalance in a dataset of diesel generator operations in Cameroon. Interpretability is achieved through SHAP (SHapley Additive exPlanations), while fairness is quantified using the Disparate Impact Ratio (DIR) across operational clusters. We further evaluate model generalization using Maximum Mean Discrepancy (MMD) to capture domain shifts between regions. Experimental results show that ensemble models consistently outperform baselines, with LightGBM achieving an F1-score of 0.99 and minimal bias across clusters (DIR \approx 0.95). SHAP analysis highlights fuel consumption rate and runtime per day as dominant predictors, providing actionable insights for operators. Our findings demonstrate that it is possible to balance performance, interpretability, and fairness in anomaly detection, paving the way for more equitable and explainable AI systems in industrial power management. {\color{black} Finally, beyond offline evaluation, we also discuss how the trained models can be deployed in practice for real-time monitoring. We show how containerized services can process in real-time, deliver low-latency predictions, and provide interpretable outputs for operators.