Are Face Embeddings Compatible Across Deep Neural Network Models?

arXiv cs.CV / 4/9/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper examines whether face embedding spaces produced by different deep neural network (DNN) models—both domain-specific and foundation models—encode identity in compatible geometric ways.
  • By modeling embeddings as point clouds, the authors test whether simple affine (low-capacity linear) mappings can align one model’s face representations to another’s.
  • Results show that these linear alignments substantially boost cross-model face identification and verification compared with using unaligned embeddings.
  • The study finds alignment behaviors that generalize across datasets and differ systematically across model families, suggesting representational convergence in how facial identity is encoded.
  • The findings have downstream implications for interoperability between biometric models, ensemble/combination strategies, and potential biometric template security considerations.

Abstract

Automated face recognition has made rapid strides over the past decade due to the unprecedented rise of deep neural network (DNN) models that can be trained for domain-specific tasks. At the same time, foundation models that are pretrained on broad vision or vision-language tasks have shown impressive generalization across diverse domains, including biometrics. This raises an important question: Do different DNN models--both domain-specific and foundation models--encode facial identity in similar ways, despite being trained on different datasets, loss functions, and architectures? In this regard, we directly analyze the geometric structure of embedding spaces imputed by different DNN models. Treating embeddings of face images as point clouds, we study whether simple affine transformations can align face representations of one model with another. Our findings reveal surprising cross-model compatibility: low-capacity linear mappings substantially improve cross-model face recognition over unaligned baselines for both face identification and verification tasks. Alignment patterns generalize across datasets and vary systematically across model families, indicating representational convergence in facial identity encoding. These findings have implications for model interoperability, ensemble design, and biometric template security.