Real-Time Branch-to-Tool Distance Estimation for Autonomous UAV Pruning: Benchmarking Five DEFOM-Stereo Variants from Simulation to Jetson Deployment

arXiv cs.CV / 3/30/2026

💬 OpinionSignals & Early TrendsModels & Research

Key Points

  • The paper targets safety-critical UAV tree pruning by estimating the metric distance from the cutting tool to thin branches in real time using a stereo depth approach.
  • It trains five DEFOM-Stereo variants on a task-specific Unreal Engine 5 synthetic dataset (5,520 stereo pairs across 115 tree instances) and deploys the resulting checkpoints to an NVIDIA Jetson Orin Super 16 GB.
  • While DEFOM-Stereo ViT-S achieves the best depth accuracy on the synthetic test set, it runs at only ~2.2 FPS on the Jetson, which is too slow for responsive closed-loop tool control.
  • The newly introduced DEFOM-PrunePlus (~21M parameters) improves the accuracy-latency trade-off, reaching ~3.3 FPS with deployable performance deemed sufficient for real-time guidance at the 2m operating range.
  • Faster lightweight variants (DEFOM-PruneStereo and DEFOM-PruneNano) meet higher frame rates but show substantially worse depth accuracy, and the authors report zero-shot results on real photos to support sim-to-real transfer for the full-capacity models.

Abstract

Autonomous tree pruning with unmanned aerial vehicles (UAVs) is a safety-critical real-world task: the onboard perception system must estimate the metric distance from a cutting tool to thin tree branches in real time so that the UAV can approach, align, and actuate the pruner without collision. We address this problem by training five variants of DEFOM-Stereo - a recent foundation-model-based stereo matcher - on a task-specific synthetic dataset and deploying the checkpoints on an NVIDIA Jetson Orin Super 16 GB. The training corpus is built in Unreal Engine 5 with a simulated ZED Mini stereo camera capturing 5,520 stereo pairs across 115 tree instances from three viewpoints at 2m distance; dense EXR depth maps provide exact, spatially complete supervision for thin branches. On the synthetic test set, DEFOM-Stereo ViT-S achieves the best depth-domain accuracy (EPE 1.74 px, D1-all 5.81%, delta-1 95.90%, depth MAE 23.40 cm) but its Jetson inference speed of ~2.2 FPS (~450 ms per frame) remains too slow for responsive closed-loop tool control. A newly introduced balanced variant, DEFOM-PrunePlus (~21M backbone, ~3.3 FPS on Jetson), offers the best deployable accuracy-speed trade-off (EPE 5.87 px, depth MAE 64.26 cm, delta-1 87.59%): its frame rate is sufficient for real-time guidance and its depth accuracy supports safe branch approach planning at the 2m operating range. The lightweight DEFOM-PruneStereo (~6.9 FPS) and DEFOM-PruneNano (~8.5 FPS) run fast but sacrifice substantial accuracy (depth MAE > 57 cm), making estimates too unreliable for safe actuation. Zero-shot inference on real photographs confirms that full-capacity models preserve branch geometry, validating the sim-to-real transfer. We conclude that DEFOM-PrunePlus provides the most practical accuracy-latency balance for onboard distance estimation, while ViT-S serves as the reference for future hardware.