A Deformable Attention-Based Detection Transformer with Cross-Scale Feature Fusion for Industrial Coil Spring Inspection
arXiv cs.CV / 3/17/2026
📰 NewsModels & Research
Key Points
- MSD-DETR introduces a structural re-parameterization strategy that decouples training-time multi-branch topology from inference-time efficiency, improving feature extraction while preserving real-time performance.
- It employs a deformable attention mechanism enabling content-adaptive spatial sampling to focus on defect-relevant regions despite morphological diversity and scale variations in coil springs.
- The approach uses cross-scale feature fusion with GSConv modules and VoVGSCSP blocks for effective multi-resolution information aggregation.
- On a real-world locomotive coil spring dataset, MSD-DETR achieves 92.4% mAP@0.5 at 98 FPS, outperforming YOLOv8 and RT-DETR while maintaining comparable speed, setting a new benchmark for industrial coil spring inspection.
Related Articles
[D] Matryoshka Representation Learning
Reddit r/MachineLearning
Two new Qwen3.5 “Neo” fine‑tunes focused on fast, efficient reasoning
Reddit r/LocalLLaMA

HKIC, Gobi Partners and HKU team up for fund backing university research start-ups
SCMP Tech
Yann LeCun’s New LeWorldModel (LeWM) Research Targets JEPA Collapse in Pixel-Based Predictive World Modeling
MarkTechPost
Streaming experts
Simon Willison's Blog