GeoRelight: Learning Joint Geometrical Relighting and Reconstruction with Flexible Multi-Modal Diffusion Transformers
arXiv cs.CV / 4/23/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- GeoRelight tackles relighting a person from a single photo by jointly estimating both 3D geometry and illumination-related appearance, addressing ambiguity that entangles geometry, intrinsic appearance, and lighting in 2D images.
- The method uses a unified Multi-Modal Diffusion Transformer (DiT) to avoid limitations of sequential pipelines and improve physical consistency by explicitly incorporating 3D geometry.
- It introduces isotropic NDC-Orthographic Depth (iNOD), a distortion-free 3D representation designed to be compatible with latent diffusion models.
- It employs a mixed-data training strategy that combines synthetic data with auto-labeled real data to improve robustness and performance.
- The paper reports that joint geometry-and-relighting yields better results than both sequential approaches and prior systems that did not leverage geometry explicitly.
Related Articles

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Trajectory Forecasts in Unknown Environments Conditioned on Grid-Based Plans
Dev.to

Why use an AI gateway at all?
Dev.to

OpenAI Just Named It Workspace Agents. We Open-Sourced Our Lark Version Six Months Ago
Dev.to

GPT Image 2 Subject-Lock Editing: A Practical Guide to input_fidelity
Dev.to