AI Navigate

TrackDeform3D: Markerless and Autonomous 3D Keypoint Tracking and Dataset Collection for Deformable Objects

arXiv cs.CV / 3/19/2026

📰 NewsTools & Practical UsageModels & Research

Key Points

  • TrackDeform3D presents an autonomous framework that uses RGB-D cameras to collect 3D datasets of deformable objects without manual annotation or motion capture setups.
  • The method identifies 3D keypoints and robustly tracks their trajectories, incorporating motion consistency constraints for temporally smooth and geometrically coherent data.
  • The approach shows consistent improvements in geometric and tracking accuracy compared to state-of-the-art methods across diverse object categories.
  • The paper provides a high-quality, large-scale dataset containing 6 deformable objects and 110 minutes of trajectory data to support downstream tasks such as dynamics modeling and motion planning.
  • By reducing data collection costs and reliance on labor-intensive labeling, TrackDeform3D aims to accelerate research and development in deformable object perception.

Abstract

Structured 3D representations such as keypoints and meshes offer compact, expressive descriptions of deformable objects, jointly capturing geometric and topological information useful for downstream tasks such as dynamics modeling and motion planning. However, robustly extracting such representations remains challenging, as current perception methods struggle to handle complex deformations. Moreover, large-scale 3D data collection remains a bottleneck: existing approaches either require prohibitive data collection efforts, such as labor-intensive annotation or expensive motion capture setups, or rely on simplifying assumptions that break down in unstructured environments. As a result, large-scale 3D datasets and benchmarks for deformable objects remain scarce. To address these challenges, this paper presents an affordable and autonomous framework for collecting 3D datasets of deformable objects using only RGB-D cameras. The proposed method identifies 3D keypoints and robustly tracks their trajectories, incorporating motion consistency constraints to produce temporally smooth and geometrically coherent data. TrackDeform3D is evaluated against several state-of-the-art tracking methods across diverse object categories and demonstrates consistent improvements in both geometric and tracking accuracy. Using this framework, this paper presents a high-quality, large-scale dataset consisting of 6 deformable objects, totaling 110 minutes of trajectory data.