AffordTissue: Dense Affordance Prediction for Tool-Action Specific Tissue Interaction
arXiv cs.CV / 4/3/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- AffordTissue is a new multimodal framework for predicting tool-action-specific safe tissue interaction regions in surgical settings, outputting dense affordance heatmaps for cholecystectomy.
- The method combines a temporal vision encoder (capturing tool motion and tissue dynamics), language conditioning (to generalize across instrument-action pairs), and a DiT-style decoder for dense affordance prediction.
- The paper introduces the first tissue affordance benchmark by curating and annotating 15,638 video clips across 103 cholecystectomy procedures, covering six tool-action pairs and four instruments.
- Experiments report substantially better dense prediction accuracy than vision-language model baselines (20.6 px ASSD vs. 60.2 px for Molmo-VLM), suggesting task-specific architectures outperform general foundation models for this dense spatial reasoning task.
- By explicitly localizing where instruments should interact safely, AffordTissue aims to improve surgical automation predictability and could enable policy guidance and early safe-stop when actions deviate from predicted zones.
Related Articles

Black Hat Asia
AI Business

Mistral raises $830M, 9fin hits unicorn status, and new Tech.eu Summit speakers unveiled
Tech.eu

ChatGPT costs $20/month. I built an alternative for $2.99.
Dev.to

OpenAI shifts to usage-based pricing for Codex in ChatGPT business plans
THE DECODER

Why I built an AI assistant that doesn't know who you are
Dev.to