H2VLR: Heterogeneous Hypergraph Vision-Language Reasoning for Few-Shot Anomaly Detection
arXiv cs.CV / 4/17/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces H2VLR, a framework for few-shot anomaly detection that leverages vision-language reasoning rather than relying on simple feature matching.
- Existing VLM-based FSAD methods are criticized for treating anomaly inference as mostly pairwise matching and for ignoring structural dependencies and global consistency.
- H2VLR reformulates FSAD as a high-order inference problem by building a heterogeneous hypergraph that jointly models visual regions and semantic concepts.
- Experiments on industrial and medical benchmarks show that H2VLR achieves state-of-the-art performance in representative few-shot anomaly detection settings.
- The authors plan to release the code after acceptance, enabling further validation and reuse by the community.



