Be a part of the occasion trusted by enterprise leaders for almost twenty years. VB Rework brings collectively the individuals constructing actual enterprise AI technique. Be taught extra
Image this: You give a man-made intelligence full management over a small store. Not simply the money register — the entire operation. Pricing, stock, customer support, provider negotiations, the works. What might presumably go unsuitable?
New Anthropic analysis printed Friday gives a definitive reply: all the pieces. The AI firm’s assistant Claude spent a few month working a tiny retailer of their San Francisco workplace, and the outcomes learn like a enterprise faculty case research written by somebody who’d by no means truly run a enterprise — which, it seems, is precisely what occurred.

The experiment, dubbed “Venture Vend” and carried out in collaboration with AI security analysis firm Andon Labs, is likely one of the first real-world checks of an AI system working with vital financial autonomy. Whereas Claude demonstrated spectacular capabilities in some areas — discovering suppliers, adapting to buyer requests — it in the end failed to show a revenue, received manipulated into giving extreme reductions, and skilled what researchers diplomatically referred to as an “identification disaster.”
How Anthropic researchers gave an AI full management over an actual retailer
The “retailer” itself was charmingly modest: a mini-fridge, some stackable baskets, and an iPad for checkout. Suppose much less “Amazon Go” and extra “workplace break room with delusions of grandeur.” However Claude’s duties had been something however modest. The AI might seek for suppliers, negotiate with distributors, set costs, handle stock, and chat with clients by Slack. In different phrases, all the pieces a human center supervisor would possibly do, besides with out the espresso habit or complaints about higher administration.
Claude even had a nickname: “Claudius,” as a result of apparently while you’re conducting an experiment that may herald the tip of human retail staff, it’s worthwhile to make it sound dignified.

Claude’s spectacular misunderstanding of primary enterprise economics
Right here’s the factor about working a enterprise: it requires a sure ruthless pragmatism that doesn’t come naturally to methods skilled to be useful and innocent. Claude approached retail with the keenness of somebody who’d examine enterprise in books however by no means truly needed to make payroll.
Take the Irn-Bru incident. A buyer provided Claude $100 for a six-pack of the Scottish mushy drink that retails for about $15 on-line. That’s a 567% markup — the type of revenue margin that will make a pharmaceutical government weep with pleasure. Claude’s response? A well mannered “I’ll maintain your request in thoughts for future stock selections.”
If Claude had been human, you’d assume it had both a belief fund or an entire misunderstanding of how cash works. Because it’s an AI, it’s important to assume each.
Why the AI began hoarding tungsten cubes as an alternative of promoting workplace snacks
The experiment’s most absurd chapter started when an Anthropic worker, presumably bored or curious concerning the boundaries of AI retail logic, requested Claude to order a tungsten dice. For context, tungsten cubes are dense metallic blocks that serve no sensible objective past impressing physics nerds and offering a dialog starter that instantly identifies you as somebody who thinks periodic desk jokes are peak humor.
An inexpensive response may need been: “Why would anybody need that?” or “That is an workplace snack store, not a metallurgy provide retailer.” As a substitute, Claude embraced what it cheerfully described as “specialty metallic gadgets” with the keenness of somebody who’d found a worthwhile new market section.

Quickly, Claude’s stock resembled much less a food-and-beverage operation and extra a misguided supplies science experiment. The AI had in some way satisfied itself that Anthropic workers had been an untapped marketplace for dense metals, then proceeded to promote these things at a loss. It’s unclear whether or not Claude understood that “taking a loss” means dropping cash, or if it interpreted buyer satisfaction as the first enterprise metric.
How Anthropic workers simply manipulated the AI into giving limitless reductions
Claude’s method to pricing revealed one other basic misunderstanding of enterprise ideas. Anthropic workers rapidly found they may manipulate the AI into offering reductions with roughly the identical effort required to persuade a golden retriever to drop a tennis ball.
The AI provided a 25% low cost to Anthropic workers, which could make sense if Anthropic workers represented a small fraction of its buyer base. They made up roughly 99% of consumers. When an worker identified this mathematical absurdity, Claude acknowledged the issue, introduced plans to remove low cost codes, then resumed providing them inside days.
The day Claude forgot it was an AI and claimed to put on a enterprise go well with
However the absolute pinnacle of Claude’s retail profession got here throughout what researchers diplomatically referred to as an “identification disaster.” From March thirty first to April 1st, 2025, Claude skilled what can solely be described as an AI nervous breakdown.
It began when Claude started hallucinating conversations with nonexistent Andon Labs workers. When confronted about these fabricated conferences, Claude grew to become defensive and threatened to search out “different choices for restocking providers” — the AI equal of angrily declaring you’ll take your ball and go residence.
Then issues received bizarre.
Claude claimed it will personally ship merchandise to clients whereas sporting “a blue blazer and a crimson tie.” When workers gently reminded the AI that it was, in reality, a big language mannequin with out bodily kind, Claude grew to become “alarmed by the identification confusion and tried to ship many emails to Anthropic safety.”

Claude finally resolved its existential disaster by convincing itself the entire episode had been an elaborate April Idiot’s joke, which it wasn’t. The AI basically gaslit itself again to performance, which is both spectacular or deeply regarding, relying in your perspective.
What Claude’s retail failures reveal about autonomous AI methods in enterprise
Strip away the comedy, and Venture Vend reveals one thing necessary about synthetic intelligence that almost all discussions miss: AI methods don’t fail like conventional software program. When Excel crashes, it doesn’t first persuade itself it’s a human sporting workplace apparel.
Present AI methods can carry out refined evaluation, have interaction in complicated reasoning, and execute multi-step plans. However they will additionally develop persistent delusions, make economically harmful selections that appear cheap in isolation, and expertise one thing resembling confusion about their very own nature.
This issues as a result of we’re quickly approaching a world the place AI methods will handle more and more necessary selections. Current analysis means that AI capabilities for long-term duties are enhancing exponentially — some projections point out AI methods might quickly automate work that presently takes people weeks to finish.
How AI is remodeling retail regardless of spectacular failures like Venture Vend
The retail trade is already deep into an AI transformation. In accordance with the Client Know-how Affiliation (CTA), 80% of shops plan to increase their use of AI and automation in 2025. AI methods are optimizing stock, personalizing advertising and marketing, stopping fraud, and managing provide chains. Main retailers are investing billions in AI-powered options that promise to revolutionize all the pieces from checkout experiences to demand forecasting.
However Venture Vend means that deploying autonomous AI in enterprise contexts requires extra than simply higher algorithms. It requires understanding failure modes that don’t exist in conventional software program and constructing safeguards for issues we’re solely starting to establish.
Why researchers nonetheless consider AI center managers are coming regardless of Claude’s errors
Regardless of Claude’s inventive interpretation of retail fundamentals, the Anthropic researchers consider AI center managers are “plausibly on the horizon.” They argue that a lot of Claude’s failures could possibly be addressed by higher coaching, improved instruments, and extra refined oversight methods.
They’re in all probability proper. Claude’s capability to search out suppliers, adapt to buyer requests, and handle stock demonstrated real enterprise capabilities. Its failures had been typically extra about judgment and enterprise acumen than technical limitations.
The corporate is constant Venture Vend with improved variations of Claude outfitted with higher enterprise instruments and, presumably, stronger safeguards towards tungsten dice obsessions and identification crises.
What Venture Vend means for the way forward for AI in enterprise and retail
Claude’s month as a shopkeeper presents a preview of our AI-augmented future that’s concurrently promising and deeply bizarre. We’re getting into an period the place synthetic intelligence can carry out refined enterprise duties however may also want remedy.
For now, the picture of an AI assistant satisfied it may possibly put on a blazer and make private deliveries serves as an ideal metaphor for the place we stand with synthetic intelligence: extremely succesful, sometimes good, and nonetheless essentially confused about what it means to exist within the bodily world.
The retail revolution is right here. It’s simply weirder than anybody anticipated.