Towards Minimal Focal Stack in Shape from Focus

arXiv cs.CV / 4/3/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Shape from Focus (SFF) is a depth reconstruction method that typically needs densely sampled, large focal stacks to estimate scene structure from focus changes.
  • The study introduces physics-based focal stack augmentation that allows SFF to work with only two input images by adding an estimated all-in-focus (AiF) image and Energy-of-Difference (EOD) maps as auxiliary cues.
  • A deep neural approach is proposed that builds a deep focus volume from the augmented stacks and iteratively refines depth using multi-scale convolutional GRUs (ConvGRUs).
  • Experiments on synthetic and real-world datasets show that the augmentation improves existing state-of-the-art SFF models and preserves comparable accuracy even with a minimal focal stack size.

Abstract

Shape from Focus (SFF) is a depth reconstruction technique that estimates scene structure from focus variations observed across a focal stack, that is, a sequence of images captured at different focus settings. A key limitation of SFF methods is their reliance on densely sampled, large focal stacks, which limits their practical applicability. In this study, we propose a focal stack augmentation that enables SFF methods to estimate depth using a reduced stack of just two images, without sacrificing precision. We introduce a simple yet effective physics-based focal stack augmentation that enriches the stack with two auxiliary cues: an all-in-focus (AiF) image estimated from two input images, and Energy-of-Difference (EOD) maps, computed as the energy of differences between the AiF and input images. Furthermore, we propose a deep network that computes a deep focus volume from the augmented focal stacks and iteratively refines depth using convolutional Gated Recurrent Units (ConvGRUs) at multiple scales. Extensive experiments on both synthetic and real-world datasets demonstrate that the proposed augmentation benefits existing state-of-the-art SFF models, enabling them to achieve comparable accuracy. The results also show that our approach maintains state-of-the-art performance with a minimal stack size.

Towards Minimal Focal Stack in Shape from Focus | AI Navigate