EX-FIQA: Leveraging Intermediate Early eXit Representations from Vision Transformers for Face Image Quality Assessment

arXiv cs.CV / 4/28/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses Face Image Quality Assessment by arguing that Vision Transformers’ intermediate-layer representations contain quality-relevant information that final-layer features alone miss.
  • It provides a comprehensive analysis across all 12 transformer blocks, showing that different depths encode distinct and complementary quality cues, reflected in attention patterns and performance differences.
  • The authors propose an early-exit and score-fusion framework that combines predictions from multiple transformer blocks using depth-weighted averaging, without architectural changes or extra training.
  • Experiments across eight benchmark datasets using four face recognition models show that the fusion approach outperforms single-exit baselines while enabling a favorable compute–performance trade-off through adaptive inference.
  • Overall, the work challenges the assumption that only deep features matter for face analysis and suggests practical deployment benefits for resource-constrained biometric systems.

Abstract

Face Image Quality Assessment is crucial for reliable face recognition systems, yet existing Vision Transformer-based approaches rely exclusively on final-layer representations, ignoring quality-relevant information captured at intermediate network depths. This paper presents the first comprehensive investigation of how intermediate representations within ViTs contribute to face quality assessment through early exit mechanisms and score fusion strategies. We systematically analyze all twelve transformer blocks of ViT-FIQA architectures, demonstrating that different depths capture distinct and complementary quality-relevant information, as evidenced by varying attention patterns and performance characteristics across network layers. We propose a score fusion framework that combines quality predictions from multiple transformer blocks without architectural modifications or additional training. Our early exit analysis reveals optimal performance-efficiency trade-offs, enabling significant computational savings while maintaining competitive performance. Through extensive evaluation across eight benchmark datasets using four FR models, we demonstrate that our fusion strategy improves upon single-exit approaches. Our proposed quality fusion approach employs depth-weighted averaging that assigns progressively higher importance to deeper transformer blocks, achieving the best quality assessment performance by effectively leveraging the hierarchical nature of feature learning in ViTs. Our work challenges the conventional wisdom that only deep features matter for face analysis, revealing that intermediate representations contain valuable information for quality assessment. The proposed framework offers practical benefits for real-world biometric systems by enabling adaptive computation based on resource constraints while maintaining competitive quality assessment capabilities.