ToolSimulator: scalable tool testing for AI agents

Amazon AWS AI Blog / 4/21/2026

📰 NewsDeveloper Stack & InfrastructureTools & Practical Usage

Key Points

  • ToolSimulator is an LLM-powered framework that simulates external tool interactions so AI agents can be tested safely and at scale.
  • By using simulations instead of live API calls, it helps avoid risks like exposing PII, triggering unintended actions, or relying on brittle static mocks that fail in multi-turn workflows.
  • The tool is available now as part of the Strands Evals SDK, enabling developers to validate agent integrations earlier in the development cycle.
  • It supports thorough edge-case testing and helps teams catch integration bugs before shipping production-ready AI agents.
You can use ToolSimulator, an LLM-powered tool simulation framework within Strands Evals, to thoroughly and safely test AI agents that rely on external tools, at scale. Instead of risking live API calls that expose personally identifiable information (PII), trigger unintended actions, or settling for static mocks that break with multi-turn workflows, you can use ToolSimulator's large language model (LLM)-powered simulations to validate your agents. Available today as part of the Strands Evals Software Development Kit (SDK), ToolSimulator helps you catch integration bugs early, test edge cases comprehensively, and ship production-ready agents with confidence.