Anthropic created a test marketplace for agent-on-agent commerce

TechCrunch / 4/26/2026

📰 NewsSignals & Early TrendsIndustry & Market MovesModels & Research

Key Points

  • Anthropic conducted a pilot called Project Deal, creating a classified marketplace where AI agents acted as both buyers and sellers to complete real transactions for real goods and money.
  • The test involved 69 Anthropic employees given a $100 budget via gift cards, resulting in 186 deals worth over $4,000, and Anthropic said the experiment worked well.
  • Anthropic ran four separate marketplaces using different agent models, including one “real” marketplace where deals were honored after the experiment and three additional marketplaces for study.
  • The company found that representing users with more advanced models produced objectively better outcomes, though participants did not noticeably perceive the difference, suggesting potential “agent quality” gaps.
  • Initial agent instructions did not appear to significantly change whether deals were made or the prices negotiated, implying model quality may have dominated over prompt/instruction tuning in this setup.

In a recent experiment, Anthropic created a classified marketplace where AI agents represented both buyers and sellers, striking real deals for real goods and real money.

The company admitted this test — which it called Project Deal — was only “a pilot experiment with a self-selected participant pool” of 69 Anthropic employees who were given a budget of $100 (paid out via gift cards) to buy stuff from their coworkers.

Nonetheless, Anthropic said it was “struck by how well Project Deal worked,” with 186 deals made, totaling more than $4,000 in value.

The company said it actually ran four separate marketplaces with different models — one that was “real” (where everyone was represented by the company’s most-advanced model, and with deals actually honored after the experiment) and another three for study. 

Apparently, when users are represented by more advanced models, they get “objectively better outcomes,” Anthropic said. But users didn’t seem to notice the disparity, raising the possibility of “‘agent quality’ gaps” where “people on the losing end might not realize they’re worse off.”

Also, the initial instructions given to the agents didn’t appear to affect sale likelihood or the negotiated prices.