Active Robotic Perception for Disease Detection and Mapping in Apple Trees

arXiv cs.RO / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes an autonomous mobile active perception system to detect and precisely map fire blight symptoms in dormant apple trees, addressing the labor and spatial-resolution limits of manual scouting.
  • It combines flash-illuminated stereo RGB sensing, real-time depth estimation, instance-level segmentation, and confidence-aware semantic 3D mapping to produce volumetric occupancy and per-voxel semantic confidence maps for growers.
  • To improve observation quality inside dense canopies, the authors evaluate three viewpoint-planning strategies within a perception-action loop, including deterministic, volumetric next-best-view, and semantic next-best-view approaches.
  • In simulated trees, the semantic planner delivers the best F1 score after 30 viewpoints, while the volumetric planner achieves the highest ROI coverage, and both planners outperform a baseline viewpoint strategy.
  • The lab experiment on fabricated trees shows strong localization performance, with the semantic planner reaching the highest final F1, serving as a precursor to future field evaluation.

Abstract

Large-scale orchard production requires timely and precise disease monitoring, yet routine manual scouting is labor-intensive and financially impractical at the scale of modern operations. As a result, disease outbreaks are often detected late and tracked at coarse spatial resolutions, typically at the orchard-block level. We present an autonomous mobile active perception system for targeted disease detection and mapping in dormant apple trees, demonstrated on one of the most devastating diseases affecting apple today -- fire blight. The system integrates flash-illuminated stereo RGB sensing, real-time depth estimation, instance-level segmentation, and confidence-aware semantic 3D mapping to achieve precise localization of disease symptoms. Semantic predictions are fused into the volumetric occupancy map representation enabling the tracking of both occupancy and per-voxel semantic confidence, building actionable spatial maps for growers. To actively refine observations within complex canopies, we evaluate three viewpoint planning strategies within a unified perception-action loop: a deterministic geometric baseline, a volumetric next-best-view planner that maximizes unknown-space reduction, and a semantic next-best-view planner that prioritizes low-confidence symptomatic regions. Experiments on a fabricated lab tree and five simulated symptomatic trees demonstrate reliable symptom localization and mapping as a precursor to a field evaluation. In simulation, the semantic planner achieves the highest F1 score (0.6106) after 30 viewpoints, while the volumetric planner achieves the highest ROI coverage (85.82\%). In the lab setting, the semantic planner attains the highest final F1 (0.9058), with both next-best-view planners substantially improving coverage over the baseline.