Continual Visual Anomaly Detection on the Edge: Benchmark and Efficient Solutions

arXiv cs.CV / 4/9/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a benchmark for visual anomaly detection (VAD) that jointly evaluates edge deployment constraints and continual learning requirements, focusing on memory footprint, inference cost, and detection performance trade-offs.
  • It presents the first comprehensive edge-and-continual-learning VAD benchmark by evaluating seven VAD models across three lightweight backbone architectures, showing that solutions tuned for one constraint can fail when both are combined.
  • It proposes Tiny-Dinomaly, a lightweight adaptation of the Dinomaly model based on DINO, achieving substantially reduced memory (13x) and compute (20x) while improving Pixel F1 by 5 percentage points.
  • It also introduces targeted efficiency-focused modifications to PatchCore and PaDiM to make them better suited to the continual learning setting.

Abstract

Visual Anomaly Detection (VAD) is a critical task for many applications including industrial inspection and healthcare. While VAD has been extensively studied, two key challenges remain largely unaddressed in conjunction: edge deployment, where computational resources are severely constrained, and continual learning, where models must adapt to evolving data distributions without forgetting previously acquired knowledge. Our benchmark provides guidance for the selection of the optimal backbone and VAD method under joint efficiency and adaptability constraints, characterizing the trade-offs between memory footprint, inference cost, and detection performance. Studying these challenges in isolation is insufficient, as methods designed for one setting make assumptions that break down when the other constraint is simultaneously imposed. In this work, we propose the first comprehensive benchmark for VAD on the edge in the continual learning scenario, evaluating seven VAD models across three lightweight backbone architectures. Furthermore, we propose Tiny-Dinomaly, a lightweight adaptation of the Dinomaly model built on the DINO foundation model that achieves 13x smaller memory footprint and 20x lower computational cost while improving Pixel F1 by 5 percentage points. Finally, we introduce targeted modifications to PatchCore and PaDiM to improve their efficiency in the continual learning setting.