See&Say: Vision Language Guided Safe Zone Detection for Autonomous Package Delivery Drones

arXiv cs.CV / 4/16/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces See&Say, a vision-language guided framework for detecting safe package drop zones for autonomous delivery drones in cluttered, dynamic urban/suburban environments.
  • It combines geometry-based safety cues (from monocular depth gradients) with semantic perception (open-vocabulary detection masks) fused into safety maps for more robust decision-making than geometry-only or segmentation-only methods.
  • A Vision-Language Model (VLM) iteratively refines hazard detection by adjusting object category prompts over time, improving reasoning during the critical final delivery phase.
  • See&Say can propose alternative candidate drop zones when the primary pad is occupied or unsafe, using the same safety reasoning pipeline.
  • Experiments on a newly curated dataset of urban delivery scenarios with moving objects and human activity show See&Say achieves better accuracy/IoU for safety map prediction and improved alternative zone selection versus baseline approaches.

Abstract

Autonomous drone delivery systems are rapidly advancing, but ensuring safe and reliable package drop-offs remains highly challenging in cluttered urban and suburban environments where accurately identifying suitable package drop zones is critical. Existing approaches typically rely on either geometry-based analysis or semantic segmentation alone, but these methods lack the integrated semantic reasoning required for robust decision-making. To address this gap, we propose See&Say, a novel framework that combines geometric safety cues with semantic perception, guided by a Vision-Language Model (VLM) for iterative refinement. The system fuses monocular depth gradients with open-vocabulary detection masks to produce safety maps, while the VLM dynamically adjusts object category prompts and refines hazard detection across time, enabling reliable reasoning under dynamic conditions during the final delivery phase. When the primary drop-pad is occupied or unsafe, the proposed See&Say also identifies alternative candidate zones for package delivery. We curated a dataset of urban delivery scenarios with moving objects and human activities to evaluate the approach. Experimental results show that See&Say outperforms all baselines, achieving the highest accuracy and IoU for safety map prediction as well as superior performance in alternative drop zone evaluation across multiple thresholds. These findings highlight the promise of VLM-guided segmentation-depth fusion for advancing safe and practical drone-based package delivery.