FCBV-Net: Category-Level Robotic Garment Smoothing via Feature-Conditioned Bimanual Value Prediction

arXiv cs.RO / 4/16/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • FCBV-Net(Feature-Conditioned bimanual Value Network)は、3Dポイントクラウドを用いてロボットによる衣類の二腕スムージングを、カテゴリー(種類)レベルで一般化できるようにする手法を提案している。
  • 事前学習済みの密な幾何特徴を凍結し、それらを条件として二腕動作の価値(value)予測を行うことで、同一カテゴリー内の衣類バリエーションに対する頑健性を高めている。
  • 学習は下流のタスク特化コンポーネントに限定し、幾何理解と価値学習(ポリシー学習)を分離することで、過学習や共同学習による性能劣化を抑える狙いがある。
  • PyFlexシミュレーション上でCLOTH3Dを用いた評価では、未見衣類に対するSteps80の効率低下が11.5%にとどまり、2D画像ベースの96.2%より大幅に改善した。

Abstract

Category-level generalization for robotic garment manipulation, such as bimanual smoothing, remains a significant hurdle due to high dimensionality, complex dynamics, and intra-category variations. Current approaches often struggle, either overfitting with concurrently learned visual features for a specific instance or, despite Category-level perceptual generalization, failing to predict the value of synergistic bimanual actions. We propose the Feature-Conditioned bimanual Value Network (FCBV-Net), operating on 3D point clouds to specifically enhance category-level policy generalization for garment smoothing. FCBV-Net conditions bimanual action value prediction on pre-trained, frozen dense geometric features, ensuring robustness to intra-category garment variations. Trainable downstream components then learn a task-specific policy using these static features. In simulated PyFlex environments using the CLOTH3D dataset, FCBV-Net demonstrated superior category-level generalization. It exhibited only an 11.5% efficiency drop (Steps80) on unseen garments compared to 96.2% for a 2D image-based baseline, and achieved 89% final coverage, outperforming an 83% coverage from a 3D correspondence-based baseline that uses identical per-point geometric features but a fixed primitive. These results highlight that the decoupling of geometric understanding from bimanual action value learning enables better category-level generalization. Code, videos, and supplementary materials are available at the project website: https://dabaspark.github.io/fcbvnet/.