Edge AI for Automotive Vulnerable Road User Safety: Deployable Detection via Knowledge Distillation

arXiv cs.CV / 4/30/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses the challenge of running accurate vulnerable road user (VRU) object detection on edge hardware under INT8 quantization constraints.
  • It proposes a knowledge distillation (KD) approach that trains a compact YOLOv8-S student (11.2M parameters) to mimic a larger YOLOv8-L teacher (43.7M parameters), targeting both compression and INT8 robustness.
  • Experiments on BDD100K using post-training INT8 quantization show that the teacher model suffers severe accuracy loss (-23% mAP) while the KD student degrades much less (-5.6% mAP).
  • The analysis indicates KD helps transfer quantization-related calibration/precision characteristics rather than merely shrinking model capacity, yielding better precision at similar recall and substantially fewer false alarms (-44% vs the collapsed teacher).
  • At INT8, the KD student even surpasses the teacher’s FP32 precision (0.748 vs. 0.718) despite being about 3.9x smaller, supporting KD as a practical requirement for safety-critical edge VRU detection.

Abstract

Deploying accurate object detection for Vulnerable Road User (VRU) safety on edge hardware requires balancing model capacity against computational constraints. Large models achieve high accuracy but fail under INT8 quantization required for edge deployment, while small models sacrifice detection performance. This paper presents a knowledge distillation (KD) framework that trains a compact YOLOv8-S student (11.2M parameters) to mimic a YOLOv8-L teacher (43.7M parameters), achieving 3.9x compression while preserving quantization robustness. We evaluate on full-scale BDD100K (70K training images) with Post-Training Quantization to INT8. The teacher suffers catastrophic degradation under INT8 (-23% mAP), while the KD student retains accuracy (-5.6% mAP). Analysis reveals that KD transfers precision calibration rather than raw detection capacity: the KD student achieves 0.748 precision versus 0.653 for direct training at INT8, a 14.5% gain at equivalent recall, reducing false alarms by 44% versus the collapsed teacher. At INT8, the KD student exceeds the teacher's FP32 precision (0.748 vs. 0.718) in a model 3.9x smaller. These findings establish knowledge distillation as a requirement for deploying accurate, safety-critical VRU detection on edge hardware.