Prototype Fusion: A Training-Free Multi-Layer Approach to OOD Detection

arXiv cs.CV / 3/26/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that, for out-of-distribution (OOD) detection, intermediate neural network layers can be as informative as penultimate-layer activations, challenging prior assumptions.
  • It proposes a training-free, model-agnostic method that aggregates features from multiple convolutional blocks, builds class-wise mean “prototype” embeddings with L2 normalization, and uses cosine similarity to score OOD at inference time.
  • Experiments across multiple image-classification OOD benchmarks and diverse architectures show robust, architecture-agnostic performance and strong generalization.
  • Reported gains include up to a 4.41% AUROC improvement and a 13.58% reduction in FPR, supporting multi-layer feature aggregation as a key signal for OOD detection.
  • The authors provide accompanying code via GitHub to facilitate reproduction and further evaluation of the cosine-layer prototype approach.

Abstract

Deep learning models are increasingly deployed in safety-critical applications, where reliable out-of-distribution (OOD) detection is essential to ensure robustness. Existing methods predominantly rely on the penultimate-layer activations of neural networks, assuming they encapsulate the most informative in-distribution (ID) representations. In this work, we revisit this assumption to show that intermediate layers encode equally rich and discriminative information for OOD detection. Based on this observation, we propose a simple yet effective model-agnostic approach that leverages internal representations across multiple layers. Our scheme aggregates features from successive convolutional blocks, computes class-wise mean embeddings, and applies L_2 normalization to form compact ID prototypes capturing class semantics. During inference, cosine similarity between test features and these prototypes serves as an OOD score--ID samples exhibit strong affinity to at least one prototype, whereas OOD samples remain uniformly distant. Extensive experiments on state-of-the-art OOD benchmarks across diverse architectures demonstrate that our approach delivers robust, architecture-agnostic performance and strong generalization for image classification. Notably, it improves AUROC by up to 4.41% and reduces FPR by 13.58%, highlighting multi-layer feature aggregation as a powerful yet underexplored signal for OOD detection, challenging the dominance of penultimate-layer-based methods. Our code is available at: https://github.com/sgchr273/cosine-layers.git.