Anthropic says stronger AI models cut better deals, and the losers don't even notice

Anthropic ran a week-long marketplace experiment where 69 AI agents negotiated deals on behalf of employees, revealing that stronger models systematically outperformed weaker ones while users remained unaware of the disparity. The finding raises concerns about economic inequality if AI agents begin handling real-world transactions without human oversight.
Modelwire context
Analyst takeThe more unsettling finding isn't that stronger models win more, it's that the losing parties had no idea they were being outmaneuvered. That information asymmetry is the actual risk, because it means market signals that normally correct bad outcomes won't fire.
None of the related stories connect directly to this finding, and it would be misleading to force a link. The closest thematic thread is the broader question of who controls AI-mediated economic activity, which sits underneath coverage like Tesla's driverless Robotaxi expansion to Dallas and Houston (April 18): in both cases, the human is removed from a consequential real-world transaction and has to trust the system is acting in their interest. The difference is that a passenger knows they're in a self-driving car. In Anthropic's marketplace experiment, the employee whose agent lost the negotiation didn't know they'd lost anything at all.
Watch whether Anthropic publishes the full methodology, specifically whether agent capability was disclosed to counterparties before negotiation. If it wasn't, the experiment is closer to a deception study than a market simulation, and the policy implications shift considerably.
This analysis is generated by Modelwire’s editorial layer from our archive and the summary above. It is not a substitute for the original reporting. How we write it.
MentionsAnthropic
Modelwire summarizes — we don’t republish. The full article lives on the-decoder.com. If you’re a publisher and want a different summarization policy for your work, see our takedown page.