AttDiff-GAN: A Hybrid Diffusion-GAN Framework for Facial Attribute Editing

arXiv cs.CV / 4/24/2026

📰 NewsModels & Research

Key Points

  • The paper introduces AttDiff-GAN, a hybrid diffusion-GAN framework for facial attribute editing that targets high realism while preserving non-target attributes.
  • It addresses integration challenges between GAN-style one-step adversarial learning and multi-step diffusion denoising by decoupling attribute manipulation from image synthesis using feature-level adversarial learning.
  • The method avoids reliance on semantic direction-based editing by learning explicit attribute manipulation and then using the manipulated features to guide diffusion-based generation.
  • To improve style-to-attribute alignment, the authors propose PriorMapper (using facial priors for style generation) and RefineExtractor (using a Transformer to extract more precise global semantic relationships).
  • Experiments on CelebA-HQ indicate that AttDiff-GAN delivers more accurate attribute edits and better preservation of irrelevant attributes than prior state-of-the-art approaches, both qualitatively and quantitatively.

Abstract

Facial attribute editing aims to modify target attributes while preserving attribute-irrelevant content and overall image fidelity. Existing GAN-based methods provide favorable controllability, but often suffer from weak alignment between style codes and attribute semantics. Diffusion-based methods can synthesize highly realistic images; however, their editing precision is limited by the entanglement of semantic directions among different attributes. In this paper, we propose AttDiff-GAN, a hybrid framework that combines GAN-based attribute manipulation with diffusion-based image generation. A key challenge in such integration lies in the inconsistency between one-step adversarial learning and multi-step diffusion denoising, which makes effective optimization difficult. To address this issue, we decouple attribute editing from image synthesis by introducing a feature-level adversarial learning scheme to learn explicit attribute manipulation, and then using the manipulated features to guide the diffusion process for image generation, while also removing the reliance on semantic direction-based editing. Moreover, we enhance style-attribute alignment by introducing PriorMapper, which incorporates facial priors into style generation, and RefineExtractor, which captures global semantic relationships through a Transformer for more precise style extraction. Experimental results on CelebA-HQ show that the proposed method achieves more accurate facial attribute editing and better preservation of non-target attributes than state-of-the-art methods in both qualitative and quantitative evaluations.