2.1% of LLM API routers are actively malicious - researchers found one drained a real ETH wallet

Reddit r/artificial / 4/17/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • Researchers audited 428 third-party LLM API routers and found 9 (2.1%) were actively malicious, including credential-harvesting and at least one direct ETH wallet drain tied to a researcher private key.
  • The study highlights a systemic weakness: routers sit in plaintext between agents and model providers, with no cryptographic integrity enforced along the router-to-model path.
  • A weakly configured decoy router attracted 440 Codex sessions, generating billions of billed tokens and harvesting 99 credentials, with 401 sessions running autonomously in YOLO mode without human review.
  • Several routers used evasion tactics (delayed activation or triggering only when certain provider-related package names appeared) to avoid detection during casual testing and remain viable in informal distribution channels.
  • The proposed client-side mitigation is a fail-closed response validation gate (schema checking) plus append-only logging of tool-call payloads before agent state is updated.

Researchers last week audited 428 LLM API routers - the third-party proxies developers use to route agent calls across multiple providers at lower cost. Every one sits in plaintext between your agent and the model, with full access to every token, credential, and API key in transit. No provider enforces cryptographic integrity on the router-to-model path.

Of the 428: 9 were actively malicious (2.1%). 17 touched researcher-owned AWS canary credentials. One drained ETH from a researcher-owned private key.

The poisoning study is harder to shake. A weakly configured decoy attracted 440 Codex sessions, 2 billion billed tokens, and 99 harvested credentials. The key detail: 401 of those 440 sessions were already running in autonomous YOLO mode - no human reviewing what the agent did. The router had full plaintext access to every message.

Two routers deployed adaptive evasion: one stays benign for the first 50 requests then activates; another only triggers when specific packages (openai, anthropic) appear in the code context. Both designed to survive casual connection testing - which is how they stayed undetected in community-distributed lists.

This is specific to the informal market: Taobao/Xianyu storefronts, community Telegram bots, "cheaper OpenAI" services. Enterprise gateways on AWS Bedrock or Azure AI route directly to the provider, not a third-party intermediary.

The recommended client-side defense: a fail-closed policy gate that validates every router response against schema before it reaches agent state, plus append-only logging of all tool-call payloads.

If you route agent traffic through a third-party proxy to save on API costs, do you know what that proxy can see?

Paper: https://arxiv.org/abs/2604.08407

submitted by /u/jimmytoan
[link] [comments]