Open, Reliable, and Collective: A Community-Driven Framework for Tool-Using AI Agents

arXiv cs.AI / 4/2/2026

📰 NewsDeveloper Stack & InfrastructureTools & Practical UsageModels & Research

Key Points

  • The paper argues that reliability issues in tool-using LLM agents come from both tool-invocation accuracy (agent deciding how/when to call tools) and intrinsic tool accuracy (the tools’ own correctness), noting prior research has focused more on the former.
  • It introduces OpenTools, a community-driven framework that standardizes tool schemas and offers plug-and-play wrappers to make tools easier to integrate across agent architectures.
  • OpenTools evaluates tools using automated test suites plus continuous monitoring, and publishes reliability reports that can evolve as tools change.
  • The authors also release a public web demo with predefined agents and tools, allowing users to run tasks and contribute test cases to improve coverage and evaluation.
  • Experiments reportedly show improved end-to-end reproducibility and task performance, with community-contributed task-specific tools yielding 6%-22% relative gains over an existing toolbox, reinforcing the role of intrinsic tool accuracy.

Abstract

Tool-integrated LLMs can retrieve, compute, and take real-world actions via external tools, but reliability remains a key bottleneck. We argue that failures stem from both tool-use accuracy (how well an agent invokes a tool) and intrinsic tool accuracy (the tool's own correctness), while most prior work emphasizes the former. We introduce OpenTools, a community-driven toolbox that standardizes tool schemas, provides lightweight plug-and-play wrappers, and evaluates tools with automated test suites and continuous monitoring. We also release a public web demo where users can run predefined agents and tools and contribute test cases, enabling reliability reports to evolve as tools change. OpenTools includes the core framework, an initial tool set, evaluation pipelines, and a contribution protocol. Experiments and evaluations show improved end-to-end reproducibility and task performance; community-contributed, higher-quality task-specific tools deliver 6%-22% relative gains over an existing toolbox across multiple agent architectures on downstream tasks and benchmarks, highlighting the importance of intrinsic tool accuracy.