The AI Pilot Graveyard
Zillow had the model. They had the data. They had years of Zestimate predictions behind them. In 2018, they launched their AI-powered iBuyer program and expanded it to 25 metropolitan areas. By the time they shut it down in November 2021, the division had lost $881 million for the year, they had cut 2,000 jobs (25% of their workforce), and watched their stock drop over 20% in a single day.
Meanwhile, their competitors Opendoor and Offerpad used similar AI and survived. The difference was not the algorithm. It was the guardrails.
McDonald’s made the same mistake with different technology. Their AI drive-thru pilot with IBM reached 100 locations in three years. It processed real orders. It scaled. And in June 2024, they shut it down. Order accuracy hovered around 85% (they needed 95% for broader rollout). The viral videos (bacon on ice cream, nine sweet teas when a customer ordered none) were symptoms, not the disease. The pilot validated that the technology could take orders. It did not validate that customers would tolerate the experience.
Meanwhile, Wendy’s used the same underlying technology (Google Cloud AI) and is still expanding. The difference: Wendy’s started with one location in Columbus, Ohio. After a year, they had 36 restaurants. They are planning 500 by end of 2025. Same technology. Different execution philosophy. The pattern repeats.
This is not a story about pilots failing to reach production. It is a story about what happens when they scale too fast.
The 95% stat
MIT research found that 95% of enterprise AI pilots fail to deliver measurable ROI. The number deserves scrutiny. The study defined success as P&L impact within six months of pilot launch, based on 52 executive interviews and 153 survey responses. Critics note this ignores efficiency gains, cost reductions, and longer-term value creation. IDC research using different methodology found $3.70 ROI for every $1 invested in GenAI.
But even if the 95% is directionally right rather than precisely right, the interesting question is why. The instinct is to read it as a failure rate. It is not. It is a design outcome.
Pilots are optimized to not reach production.
A pilot with synthetic data and friendly internal users tells you nothing about how your AI will behave when customers use it in ways you did not anticipate. A pilot that runs for six months generates board decks and justifies headcount. A pilot that “succeeds” on narrow metrics gives everyone permission to declare victory without shipping anything.
The S&P Global data confirms it: 42% of companies abandoned most of their AI initiatives in 2025, up from 17% in 2024. These are not technical failures. The models work. What fails is the organizational willingness to put AI in front of real users with real stakes.
Innovation theater
Steve Blank coined the term “innovation theater” for activities that shape culture but rarely ship: hackathons, design thinking workshops, AI labs that generate press releases but not production systems.
AI pilots are innovation theater’s most expensive production. They justify Chief AI Officer hires. They look like progress in quarterly updates. They give executives air cover for not making decisions. And they almost never ship.
The reason is structural. Innovation teams are measured on novelty: new capabilities, impressive demos, press coverage. Execution teams are measured on reliability: uptime, cost control, operational stability. When an innovation team succeeds and hands off to execution, the execution team has zero incentive to adopt it. It disrupts their metrics. The innovation dies in the handoff, not in the lab.
The tell is the timeline. If your pilot has been running for six months and nobody can name the ship date, it is theater. If the success criteria document is longer than the implementation plan, it is theater. If more people attend the steering committee than write code, it is theater.
What the 5% do differently
The companies that scale AI are not running better pilots. They are skipping pilots entirely.
They deploy to limited production immediately. They accept that some deployments will fail visibly. They build systems with guardrails that detect when conditions change and pull back automatically.
Netflix, Stripe, and Uber all deploy new ML models constantly. Not after six-month pilots. Constantly. Their pattern is called shadow deployment: the new model runs alongside production, processes the same requests, but only the current model’s predictions reach users. You get real production data before real exposure. That is not a pilot. That is continuous deployment with training wheels.
This is what separated Opendoor from Zillow. Both used AI to value homes. Both deployed to production. But the guardrails were different.
Opendoor’s system monitored the spread between predicted and actual sale prices. When the gap widened past a threshold, offers automatically tightened. The system did not need a human to notice the market was turning. It adjusted before the losses compounded.
Zillow’s executives overrode the algorithm through a practice called “offer calibration,” raising bids by thousands above model prices to hit growth targets. When the market turned, there was no automatic adjustment. The override was the strategy. By the time humans noticed the problem, Zillow owned 7,000 homes it could not sell.
The lesson is not “deploy faster.” The lesson is “deploy with automatic pullback.” A guardrail is not a dashboard someone checks. It is a trigger that fires without human intervention.
When pilots are legitimate
Regulated industries have legitimate compliance gates. Healthcare AI needs FDA clearance. Financial models need regulatory approval. These are not pilots. A compliance gate has external accountability, a clear definition of “done,” and a regulatory body that will shut you down if you skip it. A pilot has none of these.
The difference matters. MD Anderson’s IBM Watson oncology project spent four years and over $62 million before being quietly shelved. That was not regulatory compliance. That was a pilot with no kill switch and no external accountability, dressed up as medical research.
The tell is simple: if your pilot has external deadlines enforced by someone who can shut you down, it is compliance. If your pilot has been “waiting for legal review” for six months, the bottleneck is not legal.
The real problem
Pilots fail because they test the wrong things. This is different from the eval problem. Evals test whether the model works. Pilots test whether the organization will use it. Both fail for the same reason: they validate capability when they should validate adoption.
A pilot that validates whether GPT-4 can summarize documents is worthless. Of course it can. A pilot that validates whether your legal team will actually use AI summaries instead of reading the originals is the only question that matters.
Most pilots are technical solutions to organizational problems. They let engineering teams prove the technology works while avoiding the harder question of whether the organization will change how it operates.
The companies that scale AI do not ask “does this model perform well?” They ask “will this change how decisions get made?” Those are different questions. The first one can be answered in a sandbox. The second one requires production.
What to do instead
Kill pilots that have run longer than 90 days. The MIT data is clear: successful pilots transition to production in about 90 days. Failed ones take nine months or longer. The 90-day mark is the boundary between “scaling” and “pilot purgatory.” After that, sunk cost psychology takes over and the pilot becomes harder to kill than to keep funding.
Deploy to limited production with real users. Not friendly internal users. Real customers who will use your AI in ways you did not anticipate and complain when it breaks.
Build automated guardrails before you build the model. Not dashboards. Triggers. The Opendoor pattern: when predicted-vs-actual spread exceeds threshold, tighten offers automatically. No human in the loop. The decisions you make about monitoring and rollback matter more than the decisions you make about model architecture.
What I am still figuring out
Whether the “skip pilots, deploy with guardrails” advice applies to internal tools. The entire framework assumes an external user whose behavior you need to observe. For internal analytics, research tools, or back-office automation, the pilot and the limited production deployment look identical. The 90-day boundary may not apply when there is no market test. The harder question is whether internal deployments have their own version of pilot purgatory: tools that are “in use” but never change a decision. That pattern may not even be detectable without the forcing function of customer complaints.
The graveyard has two sections.
One is full of pilots that never shipped. Six-month validations that became twelve-month validations. Steering committees. Success criteria documents. Innovation theater dressed up as due diligence.
The other is full of production systems that shipped without guardrails. Zillow lost $881 million because executives overrode the algorithm and nothing automatically pulled them back. McDonald’s spent three years scaling to 100 locations before discovering customers would not tolerate 85% accuracy.
The 5% that scale are in neither section. They skipped the pilot and deployed with automatic pullback. Wendy’s started with one location. Opendoor built triggers that fired without human intervention. Netflix, Stripe, and Uber run shadow deployments that test in production without exposure.
Same technology. Different relationship with failure.