Signals: Trajectory Sampling and Triage for Agentic Interactions

arXiv cs.AI / 4/2/2026

💬 OpinionDeveloper Stack & InfrastructureSignals & Early TrendsModels & Research

Key Points

  • The paper introduces a lightweight, signal-based framework to triage and sample agentic LLM interaction trajectories using cheap, broadly applicable attributes that do not change online agent behavior.
  • Signals are organized into a taxonomy covering interaction issues (e.g., misalignment, stagnation, disengagement, satisfaction), execution problems (e.g., failure, looping), and environment conditions (e.g., exhaustion), computed without additional model calls.
  • In a controlled annotation study on the τ-bench tool-augmented agent benchmark, signal-based sampling reaches an 82% informativeness rate versus 74% for heuristic filtering and 54% for random sampling.
  • The method provides a 1.52× efficiency gain per informative trajectory and maintains advantages across different reward levels and task domains, indicating gains reflect true informativeness rather than overfocusing on obvious failures.
  • The authors argue the signals can serve as sampling infrastructure for post-deployment optimization and for constructing preference data from logged agent interactions.

Abstract

Agentic applications based on large language models increasingly rely on multi-step interaction loops involving planning, action execution, and environment feedback. While such systems are now deployed at scale, improving them post-deployment remains challenging. Agent trajectories are voluminous and non-deterministic, and reviewing each one, whether through human review or auxiliary LLMs, is slow and cost-prohibitive. We propose a lightweight, signal-based framework for triaging agentic interaction trajectories. Our approach computes cheap, broadly applicable signals from live interactions and attaches them as structured attributes for trajectory triage, identifying interactions likely to be informative without affecting online agent behavior. We organize signals into a coarse-grained taxonomy spanning interaction (misalignment, stagnation, disengagement, satisfaction), execution (failure, loop), and environment (exhaustion), designed for computation without model calls. In a controlled annotation study on \tau-bench, a widely used benchmark for tool-augmented agent evaluation, we show that signal-based sampling achieves an 82\% informativeness rate compared to 74\% for heuristic filtering and 54\% for random sampling, with a 1.52x efficiency gain per informative trajectory. The advantage is robust across reward strata and task domains, confirming that signals provide genuine per-trajectory informativeness gains rather than merely oversampling obvious failures. These results show that lightweight signals can serve as practical sampling infrastructure for agentic systems, and suggest a path toward preference data construction and post-deployment optimization.