Exploring the Limits of End-to-End Feature-Affinity Propagation for Single-Point Supervised Infrared Small Target Detection

arXiv cs.CV / 5/4/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies a minimalist approach to single-point supervised infrared small target detection by generating point-to-mask supervision online via in-batch, point-anchored feature-affinity propagation (GSACP), avoiding offline pseudo-label construction loops.
  • The authors identify an optimization bottleneck in this end-to-end design: because affinity targets are derived from the same evolving feature representation, training can become a self-referential loop that may sharpen boundaries or distort the feature space (formalized as “Self-Referential Propagation Drift”).
  • They propose a protocolized ablation strategy to isolate failure modes, including local EMA teacher decoupling, hard-background contrastive separation, and adaptive support-geometry adjustments.
  • Experiments on the SIRST3 dataset show that GSACP-Final reaches a new ultra-low false-alarm regime with 0.6674 mIoU and a 38% relative reduction in false-positive artifacts compared with PAL.
  • The work maps the performance boundaries of end-to-end in-batch feature propagation and argues it is suitable for deployment settings where suppressing false alarms is critical.

Abstract

Single-point supervised infrared small target detection (IRSTD) drastically reduces dense annotation costs. Current state-of-the-art (SOTA) methods achieve high precision by recovering mask supervision through explicit, offline pseudo-label construction, such as multi-stage active learning and physics-driven mask generation. In this paper, we study a minimalist alternative: generating point-to-mask supervision online through in-batch, point-anchored feature-affinity propagation. We instantiate this paradigm as GSACP, an end-to-end testbed that directly supervises the detector using hard-margin feature affinity gated by local image priors, entirely eliminating external label-evolution loops. This compact design, however, exposes an optimization bottleneck. Because the affinity target is generated from the same feature representation being optimized, training forms a self-referential loop. We theoretically formalize this as \emph{Self-Referential Propagation Drift}, a representation-supervision entanglement that can sharpen true boundaries or distort the feature space to satisfy its own targets. To systematically isolate these failure modes, we apply a protocolized single-variable ablation procedure spanning local EMA teacher decoupling, hard-background contrastive separation, and adaptive support geometry. On the SIRST3 dataset, GSACP-Final establishes a new ultra-low false-alarm operating regime, achieving a highly competitive 0.6674 mIoU while demonstrating a 38\% relative reduction in false-positive artifacts (\mathrm{Fa}$) compared with PAL. By systematically deconstructing the end-to-end paradigm, we map its performance boundaries and show that in-batch feature propagation provides a compact alternative for deployment scenarios where false-alarm suppression is paramount.

Exploring the Limits of End-to-End Feature-Affinity Propagation for Single-Point Supervised Infrared Small Target Detection | AI Navigate