All insights
April 15, 20267 min readAI strategyEnterpriseDelivery

How to scope an AI proof-of-concept that actually ships

Most enterprise AI POCs fail not on the model — they fail at the boundaries. Here's the scoping discipline that separates pilots that survive from pilots that die in the lab.

Roughly four out of five enterprise AI proofs-of-concept never make it into production. The failure almost never sounds like “the model wasn't accurate enough” — it sounds like “we couldn't figure out where it lives,” or “compliance had concerns we hadn't thought through,” or “the team that owns the workflow didn't ask for this.”

Those are scoping failures, not modelling failures. Here's the discipline we apply at the start of every Zianova engagement.

1. Pick a problem with a real owner

If you can't name the human who is accountable for the metric the AI is supposed to move, you don't have a project. You have a research budget. The first hour of scoping is finding that owner and getting the metric in writing.

2. Define the failure mode before the success mode

Successful AI launches think about what happens when the model is wrong before they think about what happens when it's right. Who reviews? What's the fallback path? What's the worst-case business impact of a wrong answer?

A POC that doesn't answer “what happens when this model is wrong?” is not ready to leave the lab.

3. Pre-decide the “ship vs. kill” threshold

Write down before you start: at what number does this graduate to a production roadmap, and at what number do we kill it? Every team that skips this step ends up shipping mediocre POCs because nobody wants to be the one to call it.

4. Build the evaluation harness first

Before any prompts or fine-tunes, build the eval set and the harness that scores it. If you can't measure it weekly with one command, you can't improve it deliberately. We treat evals as the first deliverable of every AI project, not the last.

5. Wire the integration sketch on day one

The most expensive surprises in AI projects come from the integration boundary — auth, data residency, downstream systems, audit logging. Build a one-page diagram of every system this feature touches before you write any model code. The diagram will tell you which compliance people you needed to call last month.

The pattern

  • Named owner + measurable metric
  • Documented failure modes and fallback paths
  • Pre-committed ship/kill threshold
  • Eval harness shipped before the model
  • Integration boundary mapped before line one

POCs that do all five graduate. POCs that skip even one usually don't. The difference between the two has very little to do with how clever the model is.

Chat on WhatsApp