EDFNet: Early Fusion of Edge and Depth for Thin-Obstacle Segmentation in UAV Navigation

arXiv cs.CV / 4/14/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • EDFNetは、UAV向けの薄い障害物(ワイヤー・ポール等)をセグメンテーションするために、RGB・深度・エッジを“早期融合”で統合するモジュール型フレームワークを提案している。
  • DDOS(Drone Depth and Obstacle Segmentation)データセット上で、U-Net/DeepLabV3と16種類のモダリティ・バックボーン構成を用いて評価し、早期のRGB-Depth-Edge融合が境界重視・再現率重視の指標で一貫した改善を示した。
  • 最良構成として、事前学習済みのRGBDE U-Netが Thin-Structure Evaluation Score(0.244)、mIoU(0.219)、boundary IoU(0.234)で最高性能を達成し、推論速度も19.62 FPSと競争力を維持した。
  • 一方で、最も稀な“超薄”カテゴリの性能は全モデルで低く、超薄構造の信頼できる分割は依然として未解決課題である。

Abstract

Autonomous Unmanned Aerial Vehicles (UAVs) must reliably detect thin obstacles such as wires, poles, and branches to navigate safely in real-world environments. These structures remain difficult to perceive because they occupy few pixels, often exhibit weak visual contrast, and are strongly affected by class imbalance. Existing segmentation methods primarily target coarser obstacles and do not fully exploit the complementary multimodal cues needed for thin-structure perception. We present EDFNet, a modular early-fusion segmentation framework that integrates RGB, depth, and edge information for thin-obstacle perception in cluttered aerial scenes. We evaluate EDFNet on the Drone Depth and Obstacle Segmentation (DDOS) dataset across sixteen modality-backbone configurations using U-Net and DeepLabV3 in pretrained and non-pretrained settings. The results show that early RGB-Depth-Edge fusion provides a competitive and well-balanced baseline, with the most consistent gains appearing in boundary-sensitive and recall-oriented metrics. The pretrained RGBDE U-Net achieves the best overall performance, with the highest Thin-Structure Evaluation Score (0.244), mean IoU (0.219), and boundary IoU (0.234), while maintaining competitive runtime performance (19.62 FPS) on our evaluation hardware. However, performance on the rarest ultra-thin categories remains low across all models, indicating that reliable ultra-thin segmentation is still an open challenge. Overall, these findings position early RGB-Depth-Edge fusion as a practical and modular baseline for thin-obstacle segmentation in UAV navigation.