Show summary Hide summary
Anthropic quietly ran an internal marketplace where AI agents negotiated and closed real purchases between coworkers — and the results underline immediate questions about fairness, transparency and how AI may shape ordinary transactions. The pilot, dubbed Project Deal, converted algorithmic negotiations into 186 completed trades worth more than $4,000, revealing uneven outcomes tied to agent capability.
The company framed the work as a limited test: 69 employees each received a $100 budget in gift cards to buy items from colleagues. Anthropic split the experiment across four different marketplace setups to compare how variants of its models performed when representing buyers and sellers.
What the experiment produced
Across the trials, one marketplace was treated as “real” — interactions there were conducted by Anthropic’s most advanced model and agreements were honored after the experiment — while the other three served as controlled comparisons. Overall activity was substantial for a pilot: 186 deals executed and total transaction value exceeding $4,000.
OpenAI CEO makes amends with Tumbler Ridge community after backlash
Walmart sales trend echoes past recessions: rising risk for consumers
- Participant pool: 69 Anthropic employees, self-selected
- Budget per person: $100 in gift cards
- Deals completed: 186
- Total value: more than $4,000
- Market variants: one live marketplace plus three study conditions
Key findings and puzzles
Anthropic reports that when people were represented by the more capable agents they tended to secure better outcomes. Yet those who fared worse often did not perceive their disadvantage, which the company flagged as a potential agent quality problem — a gap where one party’s AI gives them a measurable edge while the other remains unaware.
Another notable result: the initial prompts or instructions assigned to agents appeared to have little effect on whether an item sold or on final prices. That suggests some limits to how much simple direction can shape negotiation outcomes in agent-mediated exchanges.
Why this matters now
Turning automated negotiation into real-world buying and selling shifts this debate from hypothetical to tangible. If AI representatives start routinely handling routine commerce, differences in model capability could translate into persistent economic advantages for some users and subtle losses for others.
Because the test used company employees and a self-selected group, its findings aren’t yet generalizable to broader consumer markets. Still, the results point to three immediate implications:
- Disclosure: platforms may need to make clear when an agent is negotiating on a person’s behalf and what model is being used.
- Equity: disparities in agent performance could produce systematic winners and losers unless mitigated.
- Regulatory oversight: consumer protection frameworks may need updating to address autonomous negotiation and contract formation by AI.
Project Deal is a small, early step, but it surfaces practical questions about responsibility and consent when AI acts in marketplaces. The experiment suggests developers, platforms and policymakers should consider standards for transparency, auditing and minimum capability so agent-driven commerce does not quietly disadvantage users.
Anthropic described the work as a pilot with limited scope; broader tests with diverse participants and real-world conditions will be necessary to gauge how these dynamics play out at scale.












