PASDiff: Physics-Aware Semantic Guidance for Joint Real-world Low-Light Face Enhancement and Restoration

arXiv cs.CV / 3/27/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces PASDiff, a physics-aware semantic diffusion model for joint low-light face enhancement and restoration under real-world degradations like low illumination, blur, noise, and poor visibility.
  • PASDiff uses a training-free approach with inverse intensity weighting and Retinex-theory photometric constraints to recover natural illumination and chromaticity plausibly.
  • It adds a Style-Agnostic Structural Injection (SASI) mechanism that extracts facial structures from an off-the-shelf facial prior while filtering photometric biases, aiming to improve identity preservation.
  • The authors also build WildDark-Face, a new benchmark dataset of 700 complex low-light face images, to evaluate performance in realistic conditions.
  • Reported experiments show PASDiff outperforms existing methods, balancing natural lighting/color recovery with stronger identity consistency.

Abstract

Face images captured in real-world low light suffer multiple degradations-low illumination, blur, noise, and low visibility, etc. Existing cascaded solutions often suffer from severe error accumulation, while generic joint models lack explicit facial priors and struggle to resolve clear face structures. In this paper, we propose PASDiff, a Physics-Aware Semantic Diffusion with a training-free manner. To achieve a plausible illumination and color distribution, we leverage inverse intensity weighting and Retinex theory to introduce photometric constraints, thereby reliably recovering visibility and natural chromaticity. To faithfully reconstruct facial details, our Style-Agnostic Structural Injection (SASI) extracts structures from an off-the-shelf facial prior while filtering out its intrinsic photometric biases, seamlessly harmonizing identity features with physical constraints. Furthermore, we construct WildDark-Face, a real-world benchmark of 700 low-light facial images with complex degradations. Extensive experiments demonstrate that PASDiff significantly outperforms existing methods, achieving a superior balance among natural illumination, color recovery, and identity consistency.