AI Navigate

Geometry-Aware Semantic Reasoning for Training Free Video Anomaly Detection

arXiv cs.CV / 3/17/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • MM-VAD introduces a geometry-aware, training-free framework for video anomaly detection that treats anomaly inference as adaptive test-time inference instead of fixed feature matching.
  • It projects caption-derived scene representations into hyperbolic space to better preserve hierarchical structure and performs anomaly assessment via an adaptive question-answering process over a frozen large language model, with a learnable test-time prompt optimized by an unsupervised confidence-sparsity objective.
  • A covariance-aware Mahalanobis refinement is incorporated to stabilise cross-modal alignment while keeping backbone parameters fixed.
  • Empirically, MM-VAD achieves strong results on XD-Violence, UCF-Crime, ShanghaiTech, and UCSD Ped2, surpassing prior training-free methods.

Abstract

Training-free video anomaly detection (VAD) has recently emerged as a scalable alternative to supervised approaches, yet existing methods largely rely on static prompting and geometry-agnostic feature fusion. As a result, anomaly inference is often reduced to shallow similarity matching over Euclidean embeddings, leading to unstable predictions and limited interpretability, especially in complex or hierarchically structured scenes. We introduce MM-VAD, a geometry-aware semantic reasoning framework for training free VAD that reframes anomaly detection as adaptive test-time inference rather than fixed feature comparison. Our approach projects caption-derived scene representations into hyperbolic space to better preserve hierarchical structure and performs anomaly assessment through an adaptive question answering process over a frozen large language model. A lightweight, learnable prompt is optimised at test time using an unsupervised confidence-sparsity objective, enabling context-specific calibration without updating any backbone parameters. To further ground semantic predictions in visual evidence, we incorporate a covariance-aware Mahalanobis refinement that stabilises cross-modal alignment. Across four benchmarks, MM-VAD consistently improves over prior training-free methods, achieving 90.03% AUC on XD-Violence and 83.24%, 96.95%, and 98.81% on UCF-Crime, ShanghaiTech, and UCSD Ped2, respectively. Our results demonstrate that geometry-aware representation and adaptive semantic calibration provide a principled and effective alternative to static Euclidean matching in training-free VAD.