GenLCA: 3D Diffusion for Full-Body Avatars from In-the-Wild Videos

arXiv cs.CV / 4/9/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • GenLCA is a diffusion-based generative model that creates and edits photorealistic full-body 3D avatars from text and image inputs while keeping facial and full-body animations high-fidelity.
  • The method trains a full-body 3D diffusion model from partially observable 2D video data by using a repurposed pretrained avatar reconstruction model as an animatable 3D tokenizer, scaling training to millions of real-world videos.
  • Because real-world videos often contain only partial body observations, GenLCA introduces a visibility-aware diffusion training strategy that replaces invalid token regions with learnable tokens and applies losses only to valid regions to prevent blur/transparency artifacts.
  • A flow-based diffusion model is trained on the resulting 3D token dataset, aiming to preserve the photorealism and animatability properties of the underlying reconstruction model while enabling native 3D learning.
  • The authors report that GenLCA produces diverse, high-fidelity avatar generation and editing results and claims large performance improvements over existing approaches.

Abstract

We present GenLCA, a diffusion-based generative model for generating and editing photorealistic full-body avatars from text and image inputs. The generated avatars are faithful to the inputs, while supporting high-fidelity facial and full-body animations. The core idea is a novel paradigm that enables training a full-body 3D diffusion model from partially observable 2D data, allowing the training dataset to scale to millions of real-world videos. This scalability contributes to the superior photorealism and generalizability of GenLCA. Specifically, we scale up the dataset by repurposing a pretrained feed-forward avatar reconstruction model as an animatable 3D tokenizer, which encodes unstructured video frames into structured 3D tokens. However, most real-world videos only provide partial observations of body parts, resulting in excessive blurring or transparency artifacts in the 3D tokens. To address this, we propose a novel visibility-aware diffusion training strategy that replaces invalid regions with learnable tokens and computes losses only over valid regions. We then train a flow-based diffusion model on the token dataset, inherently maintaining the photorealism and animatability provided by the pretrained avatar reconstruction model. Our approach effectively enables the use of large-scale real-world video data to train a diffusion model natively in 3D. We demonstrate the efficacy of our method through diverse and high-fidelity generation and editing results, outperforming existing solutions by a large margin. The project page is available at https://onethousandwu.com/GenLCA-Page.