ULTRATHINK
Solutions
← Back to The Signal
Strategy December 10, 2025

Rethinking AI Maturity: From Pilots to a Modern AI Operating Model

Most enterprises have AI mandates and pilots everywhere, yet very little has made it into production. Here's why—and a framework for fixing it.

Nick Amabile
Nick Amabile
Founder & CEO

Most enterprises I talk to have the same story:

  • The CEO has an AI mandate.
  • There are pilots everywhere.
  • And yet, very little has made it into the hardened, boring, P&L-moving part of the business.

The data backs the discomfort up. BCG's latest research says only about a quarter of companies are seeing tangible value from AI, and just 4% have truly cutting-edge capabilities at scale. The rest—roughly 74%—haven't yet turned AI into measurable impact.

Other analyses put the failure rate of AI projects between 70% and 85%, once you account for initiatives that never leave the lab or get abandoned before they hit production.

At the same time, when AI does reach production, it works: McKinsey reports that 63% of companies using AI in a business unit see revenue uplift, and 44% see cost savings in that unit.

So the problem isn't that AI doesn't work. The problem is that most organizations don't have a reliable way to turn AI into systems that work.

That's why we created the Ultrathink AI Maturity Curve.

Why another AI maturity model?

If you've spent time in this space, you've seen a lot of frameworks already:

  • Google Cloud's AI Adoption Framework maps maturity across six themes—Lead, Learn, Access, Scale, Automate, Secure—mostly focused on building cloud-native capabilities.
  • Gartner's AI Maturity Model scores organizations on strategy, governance, engineering, data, and more, and shows that high-maturity orgs keep more AI projects running for three years or longer.

These are useful. But most of them share two blind spots:

  1. They treat capabilities as the destination, not as a means to business impact.
  2. They ignore the delivery & commercial model—the "Execution Gap" between strategy decks and production systems.

We built our curve because we kept seeing the same pattern:

  • Huge investments in strategy work and pilots.
  • A Trust Deficit in the consulting model (you get billed by the hour, not by the outcome).
  • Very few systems that anyone would bet a quarterly KPI on.

Our version is designed to answer one blunt question:

How reliably can you turn AI into production systems that move your P&L—again and again?

Everything else—skills, tools, data, governance—is in service of that.

What makes the Ultrathink curve different

Three big things:

1. Business-impact first, capabilities second

We don't start by asking "How many models are you running?" or "How AI-ready is your data?" We start with:

  • Which workflows drive your revenue, cost, and risk?
  • Where would AI actually change those numbers if it worked?
  • How many of those workflows are supported by reliable AI systems today?

Capabilities matter—but only as enablers of that map.

2. The modern AI application stack is explicit

Most frameworks vaguely talk about "platforms." We don't.

We model maturity against a concrete Modern AI Application Stack:

  • Foundation – compute & models
  • Data & Persistence
  • MLOps & Automation
  • Application & Logic (your workflows, your business rules)
  • Access & Presentation (UIs, APIs, integrations)

With cross-cutting security, governance, observability, and continuous evaluation.

Maturity is partly: how much of this stack is real, unified, and reusable across use cases—not how many SaaS tools you've bought.

3. We challenge the "AI-ready data or bust" myth

A lot of vendors still tell executives: "You're not AI-ready because your data isn't AI-ready." That conveniently turns every AI initiative into a five-year data program. It's also not how the leaders behave.

Even high-maturity organizations say data quality and availability are challenges—but they still manage to keep AI systems in production for years by choosing the right use cases and building robust governance and engineering around them.

Our view:

  • You don't need perfect data to start.
  • You do need a stack and a partner that can contain the mess, make the assumptions explicit, and improve data in lockstep with delivering value.

The Ultrathink AI Maturity Curve: 5 stages

We score organizations along three axes—People, Process, Technology—with business impact as the north star.

Here's the high-level view:

Stage 1 – AI Mandate, No Mechanism

You have an AI story. You don't have an AI system.

  • People: AI is on the board agenda, but nobody owns AI outcomes. A few enthusiasts play with copilots; everyone else waits for "the strategy."
  • Process: Strategy lives in decks. There is no shared path from idea → prioritized problem → production system. Governance is reactive or non-existent.
  • Technology: You have models and SaaS copilots, but no modern AI stack. No unified way to connect AI to your systems of record.

How we help here: Our Pathfinder Engagement and Synapse Cycle™ turn the vague mandate into a short, prioritized list of P&L-backed bets and a first pass at what your Axon-class stack needs to look like.

Stage 2 – Pilot Purgatory

You're rich in demos and poor in production value.

  • People: There's a central data/AI team; business units throw ideas over the wall. Leaders are interested, but no one's career depends on AI delivering real results.
  • Process: You're running lots of pilots with different vendors. Success is a demo, not adoption. There's no portfolio view and no discipline around killing weak experiments.
  • Technology: Every pilot stands up its own mini-stack. No shared platform. Monitoring, evaluation, and rollback are mostly wishful thinking.

This is where most of that 70–85% failure rate lives.

How we help here: We use Synapse Cycle and architectural review to kill most pilots and rescue a few. The survivors get rebuilt on Ultrathink Axon™, our implementation of the Modern AI Application Stack, so you end up with one foundation and many use cases—not the other way around.

Stage 3 – Connected Copilots on a Shared Platform

AI stops being a toy and starts doing real work.

  • People: There's a clear AI strategy tied to specific functions and KPIs. A small AI/platform COE exists, plus champions in ops, CX, engineering, etc.
  • Process: You have a repeatable path from idea → experiment → production candidate. Central guardrails and templates make it easy to experiment safely.
  • Technology: You've stood up a shared internal AI platform—Axon-style. Connectors into your core systems, a shared retrieval/RAG layer, early MLOps, and the first context-aware copilots that people actually use.

This is where central strategy and governance start to unlock bottom-up innovation. Domain experts can propose and trial new use cases without begging for bespoke infrastructure or one-off security reviews.

Stage 4 – Production AI Systems

AI owns real slices of core workflows with SLAs and P&L impact.

  • People: You run a hybrid org: central platform team plus embedded AI/product teams in key domains. Execs talk about AI in terms of cost, revenue, cycle time—not just innovation.
  • Process: AI initiatives are run as a portfolio with owners, value hypotheses, and kill criteria. Rollout, training, and iteration are handled via playbooks, not ad-hoc heroics.
  • Technology: The full Modern AI Application Stack is real and battle-tested. Multiple mission-critical workflows—claims, underwriting, collections, CX—have AI systems with SLAs and measurable impact.

Gartner's recent survey shows that in high-maturity organizations like this, 45% of AI projects stay in production for three years or more, and business units are four times likelier to trust and be ready to use new AI solutions (57% vs. 14% for low-maturity peers).

How we help here: This is where we fully lean into The Outcome Partnership: we run on Axon, and our economic model ties a portion of our fees to the KPIs those systems move. We don't win by billing more hours; we win when your systems move the numbers.

Stage 5 – AI-Native Operating Model

The Modern AI Stack is part of how you think and operate.

  • People: AI literacy is broad. AI-augmented experts are the norm. Central teams focus on new capabilities and safety; domains own outcomes.
  • Process: New workflows are designed assuming AI-augmented workers and agentic automation. You continuously reprioritize your AI portfolio based on real impact.
  • Technology: An Axon-class stack is the default runtime for new workflows. Agentic and semi-autonomous systems handle a meaningful share of decisions; humans set objectives and guardrails.

Only a small fraction of companies are here. BCG's latest numbers suggest 4–5% of organizations really qualify as AI leaders; they're the ones seeing significantly higher revenue growth and shareholder returns from AI.

What's next: Take the self-assessment

Behind this curve, we've built a simple scoring rubric. For each stage, you rate your People, Process, and Technology on a 1–5 scale with concrete statements. The result is a clear picture of where you really are today—and what has to change to climb a stage.

Take the AI Maturity Assessment →

It takes about 5 minutes. You'll get your stage, your dimension breakdown, and recommendations for your "next best move."

If you're looking at your current AI portfolio and seeing more slideware than systems, you're not alone. The good news is: you don't have to be perfect, and you don't have to start over.

You just need a clear view of where you are on the curve, a pragmatic plan for the next rung up, and a partner who's willing to put real skin in the game.

Take the assessment now to find out where you stand.

How Ultrathink Helps You Move Up the Curve

Pathfinder Engagement

A focused, time-boxed engagement to align leadership on where AI can actually move the P&L.

Synapse Cycle™ + Axon™

A repeatable path from ambiguity to production, built on a Modern AI Application Stack.

Outcome Partnership

Our economics are tied to the outcomes we deliver together, not hours logged.

Ready to Close the Execution Gap?

Take the next step from insight to action.

No sales pitches. No buzzwords. Just a straightforward discussion about your challenges.