AI Navigate

AI-Driven Predictive Maintenance with Real-Time Contextual Data Fusion for Connected Vehicles: A Multi-Dataset Evaluation

arXiv cs.LG / 3/17/2026

💬 OpinionDeveloper Stack & InfrastructureSignals & Early TrendsModels & Research

Key Points

  • It proposes a simulation-validated proof-of-concept framework for V2X-augmented predictive maintenance that fuses on-board sensor data with external contextual signals via V2X and third-party APIs, with inference performed at the vehicle edge.
  • The study finds that incorporating V2X contextual features yields a 2.6-point F1 gain, and removing full context reduces macro F1 from 0.855 to 0.807.
  • On the AI4I 2020 industrial failure dataset, LightGBM achieves AUC-ROC 0.973 under 5-fold CV with SMOTE limited to training folds.
  • SHAP analysis shows V2X and engineered interaction features rank among the top predictors, and edge inference reduces latency from 3.5s to under 1s compared with cloud-only processing, though field validation on instrumented vehicles is needed.

Abstract

Most vehicle predictive maintenance systems rely exclusively on internal diagnostic signals and are validated on deterministic synthetic data, limiting the credibility of reported metrics. This paper presents a simulation-validated proof-of-concept framework for V2X-augmented predictive maintenance, integrating on-board sensor streams with external contextual signals -- road quality, weather, traffic density, and driver behaviour -- acquired via V2X communication and third-party APIs, with inference at the vehicle edge. Field validation on instrumented vehicles is identified as the required next step. Three experiments address common shortcomings of prior work. A feature group ablation study shows that V2X contextual features contribute a 2.6-point F1 gain, with full context removal reducing macro F1 from 0.855 to 0.807. On the AI4I 2020 real-world industrial failure dataset (10,000 samples, five failure modes), LightGBM achieves AUC-ROC of 0.973 under 5-fold stratified CV with SMOTE confined to training folds. A noise sensitivity analysis shows macro F1 remains above 0.88 under low noise and degrades to 0.74 under very high noise. SHAP analysis confirms that V2X and engineered interaction features rank among the top 15 predictors. Edge inference is estimated to reduce latency from 3.5s to under 1.0s versus cloud-only processing.