A thesis on the Applied AI Firm, the Operational Leakage Map, and the decade that will decide which companies compound and which disappear. Fully sourced. Intellectually honest about its own weak points.
Executive summary
The research is now unambiguous. Between 42 and 95 percent of enterprise AI initiatives fail to produce material value, depending on how the question is framed. MIT's NANDA project put the number at 95 percent. S&P Global Market Intelligence reports the abandonment rate rose from 17 percent in 2024 to 42 percent in 2025. BCG finds that only 25 percent of firms generate meaningful value from AI, while the other 75 percent spend, iterate, and stall.
These numbers describe the same phenomenon. There is a gap between AI capability and AI value capture in operating businesses. We call it the Implementation Gap. This paper explains what it is, why traditional delivery models cannot close it, and what kind of firm can.
1. The Implementation Gap is real and measurable
The gap between the frontier of AI capability and the rate at which commercial businesses capture value from AI has widened, not narrowed, over the last 24 months. This is the opposite of what most stakeholders expected.
The evidence:
- MIT NANDA, 2025: 95 percent of generative AI pilots at large enterprises fail to deliver measurable P&L impact.
- S&P Global Market Intelligence, 2025: 42 percent of U.S. companies abandoned most of their AI initiatives in 2025, up from 17 percent in 2024.
- BCG, Build for the Future 2025: 60 percent of executives report no material value from their AI programs. Only 25 percent of firms in the study generate meaningful value.
- IDC / Lenovo, 2025: for every 33 AI proofs-of-concept launched, four reach production. An 88 percent POC failure rate.
- McKinsey State of AI, 2024 and 2025: adoption is up, but EBIT impact remains concentrated in a small minority of companies. The median company shows less than a 1 percent EBIT lift from AI.
The signal is consistent across independent studies and methodologies. Capability is compounding. Capture is not.
2. Why traditional delivery models cannot close the gap
The Implementation Gap is not a model problem. It is a delivery problem. The dominant delivery models were built for a different kind of work.
Big consulting
Deck-heavy, strategy-forward, thin on owned engineering. Great at framing and sequencing. Structurally incentivized to sell phases of work, not outcomes. The average enterprise AI strategy engagement produces a detailed roadmap and a POC that dies inside 90 days because nobody owns running it.
Agencies and dev shops
Good at building what a client specifies. Not good at diagnosing what the client should have specified. They can hand off a working pipeline. They cannot hand off business judgment. The result is a lot of well-built systems that automate the wrong process.
SaaS vendors
Sell a product. Every product is horizontal by design. The last mile of integration, data hygiene, change management, and workflow redesign is labeled "professional services" and priced as overhead. That last mile is where the value lives. Every study of AI outcomes points to the same conclusion: the model is 10 percent of the work.
Internal build
Some firms have the talent to build in-house. Most do not. Hiring a head of AI at mid-market compensation, in a market where frontier labs pay seven figures for senior engineers, is not a realistic play. For the 10,000 U.S. companies in the $10M–$500M revenue band, internal build is structurally out of reach.
Each model captures a slice. None captures the full stack of diagnosis, build, and operation. The gap is the stack.
3. The Operational Leakage Map
Our framing for where AI value actually lives in operating businesses is the Operational Leakage Map. It is deliberately horizontal, because the leakage pattern is horizontal.
Five categories of operational leakage show up on the P&L of nearly every mid-market commercial business we have examined:
- Revenue leakage. Quotes that do not close on time. Proposals that sit. Intake that drops leads. Follow-up that does not happen. Dollars walk out the door because nobody is routing them.
- Labor leakage. Senior people doing junior-people work. Manual extraction, manual reconciliation, manual routing. Talent spent on the part of the job that does not require judgment.
- Process leakage. Swivel-chair work between systems. Data re-entered three times. Status calls that exist because nothing is instrumented. The tax of running an operation without a spine.
- Risk leakage. Compliance work that happens by memory. Reviews that happen at the end, not at the edge. Liability that accumulates quietly and surfaces expensively.
- Decision leakage. Decisions made late, or with the wrong data, because the system that would support them does not exist. Forecasts that are stale. Pricing that is lagging. Capacity that is misallocated.
Each category has recognizable patterns, recognizable instrumentation, and a recognizable dollar range. The Operational Leakage Index (our quarterly benchmark, launching Q3 2026) will publish those ranges by sector.
4. The Applied AI Firm
The firm that can close the Implementation Gap for the mid-market does three things, in sequence, for one client at a time:
Diagnose
Walk into the operation. Instrument the leakage. Produce a written artifact that dollarizes each category and ranks the interventions by return and by time-to-value. This artifact is the deliverable a client should expect before any build work starts. It is the opposite of a deck.
Build
Own the engineering end to end. Not "oversee a vendor." Not "architect the solution." Build it. Integrations, pipelines, models, interfaces, monitoring. The firm that builds the system is the firm that understands why it works.
Operate
Run the system until the client's team can. In practice, this is the piece every other delivery model skips. Most AI systems fail at day 90 not because the model drifts, but because nobody owned the operational discipline of keeping the integrations current, the prompts updated, the feedback loop closed, and the workflow adjusted as the business changes.
Diagnose. Build. Operate. That is the shape of the Applied AI Firm. It is not a consultancy, not an agency, not a product, not a staffing firm. It is a fourth thing.
5. Why this shape compounds
The Applied AI Firm compounds because the diagnostic, the build, and the operation share the same instrumentation. Every engagement produces structured data about where leakage lives, at what magnitude, with what fix, at what cost, and with what outcome. That structured data is the flywheel.
Over time, the firm gets better at:
- Recognizing leakage patterns from a half-hour conversation.
- Pricing interventions against a real distribution of outcomes, not a fantasy.
- Shipping the second version of a system faster than the first, because the patterns repeat.
- Publishing benchmarks the industry does not have, because nobody else is instrumented to collect them.
The writing, the index, and the work feed each other. The firm that compounds is the firm that publishes the evidence.
6. What this thesis gets wrong, honestly
We will not pretend this framing is complete. A few weak points we are willing to name:
- Enterprise-scale work is a different animal. The Applied AI Firm shape described here is calibrated for mid-market commercial businesses in roughly the $10M–$500M revenue band. Above that, organizational politics, procurement, and regulatory complexity change the economics meaningfully. We do not claim this framing generalizes.
- Some leakage is not AI-addressable. Strategy problems, product-market-fit problems, and leadership problems cannot be instrumented away. We decline engagements where the underlying issue is not an operational issue.
- The data on AI outcomes is still young. The highest-quality studies we cite are 12–18 months old. The next two years will either confirm the pattern or complicate it. We will update this thesis as the evidence does.
- Our own sample is biased. We see the engagements that reach us. We do not see the ones that never call. Our published benchmarks will be transparent about this.
7. What happens next
Over the decade between 2026 and 2035, the gap between companies that implement AI well and companies that do not will translate directly into differences in margin, growth, and valuation. The firms that close the Implementation Gap early will compound. The firms that do not will be acquired, automated around, or disappear.
The Applied AI Firm is not the only way to close the gap. It is the way we are betting on. The structure of this firm — diagnostic first, engineering-led, operating inside the client — is calibrated for the shape of the problem as we actually see it, week after week, in the mid-market.
This paper is the first long-form argument. The Operational Leakage Index, launching Q3 2026, will be the first dataset. The engagements we take in 2026 and 2027 will be the first proof points. We expect to be wrong about some of this, and we will publish the corrections when we are.
If your business has leakage you can see on the P&L and an executive on the inside who owns the change, we should probably talk.
Sources
- MIT NANDA project, State of AI in Business 2025. (2025)
- S&P Global Market Intelligence, AI initiative abandonment survey. (2025) Link
- BCG, Build for the Future 2025: Are You Generating Value from AI? Link
- IDC / Lenovo, AI POC to production study. (2025)
- McKinsey, The State of AI. (2024 and 2025 editions)
- Gartner, Hype Cycle for Artificial Intelligence. (2024 and 2025 editions)
