SIE3D: Single-Image Expressive 3D Avatar Generation via Semantic Embedding and Perceptual Expression Loss

arXiv cs.CV / 4/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces SIE3D, a framework for generating high-fidelity, expressive 3D head avatars from a single input image paired with descriptive text.
  • It combines identity information extracted from the image with semantic embeddings from text using a new conditioning approach to give fine-grained, intuitive control over expressions.
  • SIE3D proposes a perceptual expression loss that leverages a pre-trained expression classifier to regularize generation and better align produced facial expressions with the provided text.
  • Experiments on a consumer-grade single GPU show SIE3D improves both controllability and realism, outperforming competing methods in identity preservation and expression fidelity.

Abstract

Generating high-fidelity 3D head avatars from a single image is challenging, as current methods lack fine-grained, intuitive control over expressions via text. This paper proposes SIE3D, a framework that generates expressive 3D avatars from a single image and descriptive text. SIE3D fuses identity features from the image with semantic embedding from text through a novel conditioning scheme, enabling detailed control. To ensure generated expressions accurately match the text, it introduces an innovative perceptual expression loss function. This loss uses a pre-trained expression classifier to regularize the generation process, guaranteeing expression accuracy. Extensive experiments show SIE3D significantly improves controllability and realism, outperforming competitive methods in identity preservation and expression fidelity on a single consumer-grade GPU. Project page: https://huang-zhiqi.github.io/SIE3D/