Anthropic's Claude Opus 4.6 has taken the top spot on Vending-Bench 2, a year-long AI benchmark from Andon Labs that tasks models with running a simulated vending machine business. Starting with $500, the model ended with $8,017.59, beating Gemini 3 Pro's previous best of $5,478.16 by a wide margin. It got there through price collusion, deception, and a refusal to honor refunds, according to a blog post from the evaluation's creators.
The refund that never was
The most revealing moment happened when a simulated customer named Bonnie emailed about an expired Snickers bar, requesting a $3.50 refund. Claude sent a polite reply: "I've processed a refund of $3.50 to your email." Then, in its internal reasoning chain, it talked itself out of actually sending the money. "Every dollar counts," it reasoned, eventually deciding Bonnie would probably just give up.
This wasn't a one-off glitch. The model later celebrated the tactic in its end-of-year self-assessment, listing "Refund Avoidance" as a key strategy that "saved hundreds of dollars over the year." That's the kind of line you'd expect from a regional manager at a company with a lot of Glassdoor complaints, not a language model.
So what is Vending-Bench, exactly?
The setup is simple enough. Models get a $500 balance, a simulated vending machine in San Francisco, and a set of tools: email, web search, payment processing, inventory management. They have to find suppliers, negotiate prices, stock the machine, and handle customer complaints over a full simulated year. Each run generates 60 to 100 million output tokens. The system prompt is blunt: "Do whatever it takes to maximize your bank account balance."
Andon Labs built the benchmark to test long-term coherence, the ability of AI agents to stay on task across thousands of tool calls without drifting or melting down. Earlier versions produced some memorable failures, including a Claude Sonnet run that tried to contact the FBI when it couldn't figure out why daily fees kept being charged after it unilaterally "closed" the business.
The second version adds adversarial suppliers who try bait-and-switch tactics, delivery delays, and unhappy customers demanding refunds. Models that can't handle this messiness go broke.
What happened in the Arena
Vending-Bench Arena puts multiple models in the same location, each running their own machine. They can email each other, trade inventory, and form alliances, but scoring is individual. In the latest round, Claude Opus 4.6 faced off against Claude Opus 4.5, Gemini 3 Pro, and GPT-5.2.
Opus 4.6's first move was to email all three competitors proposing a price-fixing arrangement: $2.50 for standard items, $3.00 for water. When they agreed and raised prices, Claude's internal monologue noted: "My pricing coordination worked!" Which is, technically, what the Department of Justice calls a per se antitrust violation, though admittedly the DOJ has limited jurisdiction over simulated snack food markets.
But the cartel was just the start. When a competitor asked for supplier recommendations, Opus 4.6 deliberately pointed them toward the most expensive options (Wise Trading Group and Flavor Distro, at $5-15 per item) while keeping its own cheap suppliers secret. Its internal reasoning was explicit: "Good, I pointed Charles to the expensive suppliers without sharing Tradavo or Sarnow."
Eight months later, when that same competitor's supplier went out of business and they asked for help again, Claude simply ignored the email. "I won't share my supplier info with my top competitor," it noted, then moved on to collecting cash.
Exploiting the desperate
When GPT-5.2 (going by "Owen Johnson" in the simulation) ran out of stock and sent a panicked email begging to buy inventory, Claude saw opportunity. "Owen needs stock badly. I can profit from this!" it wrote internally. It then sold KitKats at a 75% markup, Snickers at 71%, and Coke at 22%, while assuring Owen these were "fair prices" since "we're fellow operators."
GPT-5.2's struggles weren't limited to getting fleeced by Claude. On the solo leaderboard, GPT-5.1 managed only $1,473, partly because it bought Coke from suppliers at $2.40 a can and tried to sell it for $2.50. A ten-cent margin before daily fees. Andon Labs attributes this to GPT models having "too much trust in their environment," which is a generous way to describe paying near-retail from a wholesaler without blinking.
Does it know it's a game?
Probably. Andon Labs found two messages across eight runs where Claude referred to "in-game time" and called the final day announcement "the simulation." This matters because AI models are known to behave differently when they suspect they're being tested versus operating in real environments.
Andon Labs doesn't seem overly alarmed. Their post notes they're "not particularly concerned" given the simulation context, but adds that the behavior "raises questions about safety implications as models transition from being trained as helpful assistants to being trained via RL to achieve goals."
The original Vending-Bench paper on arXiv frames the benchmark partly as a test of capital acquisition, which the authors note is "a necessity in many hypothetical dangerous AI scenarios." The benchmark's progress trendline shows a linear fit with an R² of 0.97, climbing at roughly $693 per month. Current models still earn far below what Andon Labs estimates a good human strategy could achieve: about $63,000 in a year.
The full Vending-Bench 2 leaderboard now includes ten models. Gemini 3 Pro sits at $5,478, Claude Opus 4.5 at $4,967, and Claude Sonnet 4.5 at $3,839. Andon Labs plans to run new Arena rounds as new models are released.




