Tech10
Back to blog

Why 95% of AI Agent Pilots Fail in Production (And How to Be in the 5%)

Why 95 Percent Ai Agent Pilots Fail In Production
AI AgentsMar 31, 20266 min readDoreid Haddad

The 95% figure comes from MIT's NANDA project, which tracks enterprise generative AI deployments. Cleanlab's 2025 survey reaches the same neighborhood from a different angle: out of 1,837 organizations they polled, only 95 had AI agents live in production. The Berkeley/Stanford/IBM "Measuring Agents in Production" study doesn't quote a failure rate but documents what the successful 5% does differently across 26 industries and 306 practitioners.

The convergence across three independent data sources is the useful part. The failure isn't random and the success isn't mysterious. The 5% does seven specific things differently. This article is each of the seven failure modes the data identifies, the fix, and a concrete signal you can use to spot it in your own project before it kills you.

Failure mode 1: Building an agent for a workflow nobody runs

The MAP study found that 73% of production teams cite "increasing speed of task completion over the previous non-agentic system" as their reason for building. The implicit prerequisite: there was a previous system or human process. Teams that build agents for workflows nobody currently does manually don't have a baseline to beat or a savings to measure.

Signal you have this problem: when asked "what happens today without the agent," the answer is theoretical. "Well, in principle, someone could…"

Fix: Pick a workflow your team actually runs every week. Document the human time it consumes. The agent's job is to reduce that number.

Failure mode 2: No eval set, only vibes

Across Cleanlab's data and the MAP study, the dividing line between production-ready and pilot-stalled agents is whether the team has a held-out evaluation set they grade on a cadence. Without one, every prompt change is a guess and every regression is a surprise.

Signal: when asked "is the agent getting better or worse this month," the team can't answer with a number.

Fix: Build a 50-example eval set from real production cases before you tune the prompt. Grade it weekly. Add examples whenever the agent surprises you.

Failure mode 3: Confidence without thresholds

Agents that take action whenever they decide to take action — without a confidence threshold and a human review path for low-confidence cases — produce the embarrassing failures that get pilots cancelled. The Cleanlab respondent who described "moving from LangChain to Azure in two months, only to consider moving back again" was talking about systems where unbounded autonomy created reputational damage faster than the team could rebuild trust.

Signal: the agent is configured to act on every input it sees.

Fix: Implement a confidence threshold. Below the threshold, route to human review. Tune the threshold every two weeks based on grading the bottom 5% of auto-handled cases.

Failure mode 4: Skipped output validation

Anthropic's Building Effective Agents describes this implicitly when it warns about frameworks that "obscure the underlying prompts and responses." Without strict schema validation at every boundary, malformed agent outputs slip into downstream systems and create the worst class of failure: technically running, quietly wrong.

Signal: the orchestrator passes whatever the model returns to the next step without parsing it through a typed schema.

Fix: Pydantic, Zod, or JSON Schema validation at every output. Failed validation routes to a human queue with the raw output attached. Don't pass malformed objects downstream.

Failure mode 5: Tool sprawl

The Berkeley Function-Calling Leaderboard has shown for two years that tool-calling accuracy degrades past about a dozen tools. The MAP study's finding that 68% of production agents execute fewer than 10 steps before human intervention is partly a story about how few action types most workflows actually need. Production agents that ship typically have three to five tools.

Signal: the agent has more than ten tools, or the team can't quickly explain which tool handles which case.

Fix: Cut to the three tools the workflow genuinely needs (a read, a write, a decide-and-route is the canonical set). Add tools only when production logs show a specific gap.

Failure mode 6: Always-on without observability

Agents running 24/7 without per-trace logging, eval-pass-rate dashboards, and alerting on failure modes are agents whose problems get discovered by customers. Cleanlab's data shows less than one in three teams are satisfied with their observability tooling, and the gap correlates with which deployments hit retention problems three months in.

Signal: the team can't pull up a specific past agent run and explain what it did, or doesn't know whether eval pass rate has changed in the last 30 days.

Fix: Per-run tracing (every model call, every tool call, every retry). Eval pipeline running daily or per-change. Cost tracking per task. Alerts on iteration count above cap, tool error rate above threshold, schema validation failure rate above 1%, and eval pass-rate drop above 3 percentage points.

Failure mode 7: Stack churn nobody planned for

Cleanlab found that 70% of regulated enterprises rebuild their entire AI agent stack every three months or faster. Teams who didn't plan for this — who built a v1 they expected to maintain for years — find themselves painted into corners when the framework, the model, or the integration pattern needs to change. The teams who did plan for it built modular systems that absorb churn without rewrites.

Signal: the agent's core orchestration logic is so deeply tied to a specific framework that swapping the framework would require rebuilding most of the system.

Fix: Treat the agent as a backend service with clear interfaces. Model API behind an interface. Orchestration behind an interface. State storage behind an interface. Each of these can be swapped independently when the stack churns. The Anthropic recommendation to "start by using LLM APIs directly" is partly about preserving this flexibility.

What the 5% does differently

The seven fixes above aren't a list of best practices to feel good about. They're the actual differences between the 5% who ship and the 95% who don't, drawn from independent surveys that found the same patterns.

The pattern under all of them: the 5% treats AI agents as production software, not as experiments. Production software has SLAs, evals, schemas, observability, and a runbook. Production software has a planned upgrade path. Production software is owned by an engineering team with on-call coverage. Most failed pilots wanted the upside of agents without the operational discipline. The successful ones built the discipline first.

This is also why the 5% doesn't grow as quickly as people expected. The upside of agents is real — the MAP study found broad evidence of productivity gains across 26 domains. The investment to capture the upside is also real, and it's mostly in operational discipline rather than model selection. Teams who underestimate the second part stay in the 95%.

If you're sizing an AI agent project right now, the honest framing is: the model is roughly 10-20% of total project cost and 5-10% of why projects succeed or fail. The other 80-90% lives in the seven failure modes above. Plan for that and you're already most of the way to the 5%.

Frequently Asked Questions

Where does the 95% failure-rate figure come from?

MIT's NANDA project, which tracks enterprise generative AI deployments. The figure refers specifically to AI agent pilots that fail to reach production with measurable business value. Cleanlab's separate 2025 survey of 1,837 enterprises found that only 95 of them had agents live in production — roughly the same 5% rate from a different angle.

What does 'fail' actually mean in the 95% figure?

In NANDA's tracking, failure means the pilot was discontinued, never moved to production, or moved to production but didn't deliver measurable business outcomes. It does not necessarily mean the model didn't work — most failures are workflow, data, or operational, not model quality.

What's the single most predictive factor for being in the 5%?

Eval discipline. Across both the MAP study (74% of successful production agents rely primarily on human evaluation) and Cleanlab's data (63% of production teams plan to invest more in evaluation), the pattern is consistent: teams who measure their agent against held-out examples ship and stay shipped. Teams who don't, drift and stall.

Sources
Doreid Haddad
Written byDoreid Haddad

Founder, Tech10

Doreid Haddad is the founder of Tech10. He has spent over a decade designing AI systems, marketing automation, and digital transformation strategies for global enterprise companies. His work focuses on building systems that actually work in production, not just in demos. Based in Rome.

Read more about Doreid

Keep reading