Transparent Fragments Contour Estimation via Visual-Tactile Fusion for Autonomous Reassembly

arXiv cs.CV / 3/24/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • 提案論文は、透明な破片の輪郭推定を目的に、視覚と触覚(Gelsight Mini)の情報を融合する汎用フレームワークを提示しています。
  • 新規データセット「TransFrag27K」と合成データ生成パイプラインを構築し、多様な透明物体の多シーン破片を扱えるようにしています。
  • 位置検出・セグメンテーション用の視覚ネットワーク「TransFragNet」を用い、二指グリッパで触覚情報(側縁の再構成)を取得して輪郭推定と分類精度を高めます。
  • さらに、輪郭マッチングのための多次元類似度指標と、再組立アルゴリズムを組み合わせ、再現可能な評価ベンチマークを提供しています。
  • 実環境での検証結果に加え、データセットとコードが公開されている点により、自律的な再組立研究の起点として活用しやすい内容です。

Abstract

The contour estimation of transparent fragments is very important for autonomous reassembly, especially in the fields of precision optical instrument repair, cultural relic restoration, and identification of other precious device broken accidents. Different from general intact transparent objects, the contour estimation of transparent fragments face greater challenges due to strict optical properties, irregular shapes and edges. To address this issue, a general transparent fragments contour estimation framework based on visual-tactile fusion is proposed in this paper. First, we construct the transparent fragment dataset named TransFrag27K, which includes a multiscene synthetic data of broken fragments from multiple types of transparent objects, and a scalable synthetic data generation pipeline. Secondly, we propose a visual grasping position detection network named TransFragNet to identify, locate and segment the sampling grasping position. And, we use a two-finger gripper with Gelsight Mini sensors to obtain reconstructed tactile information of the lateral edge of the fragments. By fusing this tactile information with visual cues, a visual-tactile fusion material classifier is proposed. Inspired by the way humans estimate a fragment's contour combining vision and touch, we introduce a general transparent fragment contour estimation framework based on visual-tactile fusion, demonstrates strong performance in real-world validation. Finally, a multi-dimensional similarity metrics based contour matching and reassembly algorithm is proposed, providing a reproducible benchmark for evaluating visual-tactile contour estimation and fragment reassembly. The experimental results demonstrate the validity of the proposed framework. The dataset and codes are available at https://github.com/Keithllin/Transparent-Fragments-Contour-Estimation.