Assistance Without Interruption: A Benchmark and LLM-based Framework for Non-Intrusive Human-Robot Assistance

arXiv cs.RO / 5/5/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper formalizes non-intrusive human-robot assistance as a distinct paradigm in human-robot interaction, emphasizing proactive support while strictly avoiding interruptions.
  • Instead of relying on direct commands or negotiation, it treats the human’s plan as the primary process and frames assistance as a joint decision about when to act and what actions to take.
  • It introduces NIABench, a simulation benchmark, along with task-specific metrics to systematically evaluate how well non-intrusive assistance is achieved.
  • The proposed hybrid framework combines an LLM with a scoring model that uses semantic retrieval to narrow candidate actions and a ranker to evaluate human-step/robot-action pairs for timing and dependency reasoning.
  • Experiments on both NIABench and real-world scenarios indicate the approach can reduce human effort while maintaining task effectiveness.

Abstract

Human-robot interaction (HRI) has long studied how agents and people coordinate to achieve shared goals. In this work, we formalize and benchmark the non-intrusive assistance as an independent paradigm of HRI, where a robot proactively supports a human's ongoing multi-step activities while strictly avoiding interruptions. Unlike conventional HRI tasks that rely on direct commands, explicit negotiation, or proactive interventions based on user habits and history, our task treats the human's plan as the primary process and formulates assistance as a joint decision over when to act and what to do. To systematically evaluate this problem, we establish a simulation benchmark, NIABench, along with new metrics tailored to the non-intrusive assistance task. We further propose a hybrid architecture that integrates an LLM with a scoring model. The scoring model first applies semantic retrieval to prune large candidate action sets, and then a ranker evaluates human-step and robot-action pairs, enabling reasoning over timing and cross-step dependencies. Comprehensive experiments on both NIABench and real-world scenarios demonstrate that our method achieves proactive, non-intrusive assistance that reduces human effort while preserving task effectiveness.