The most common failure mode in AI adoption isn't a technology problem. It's an assessment problem. Organisations look at a workflow, decide it should be automated, and hand it to someone to build. They skip the part where you actually understand what the workflow does, how it really runs, and whether it's worth automating in its current form.

The Assess stage in the AGENTIC Framework exists because I watched this happen. Twice. At a marine conservation non-profit, a monthly reporting workflow had four data sources, not the three everyone thought. At a venture studio, an approval process had five undocumented rules living entirely in one person's head. In both cases, jumping to automation would have encoded those gaps into a system and scaled them.

AI workflow assessment is the process of understanding a workflow deeply enough to specify it for an agent. That means mapping it as it actually runs, surfacing the knowledge that lives in people's heads, fixing what's broken, and translating all of it into a specification that's both human-readable and machine-executable.


Why most AI adoption fails before it starts

The failure pattern is consistent. Someone identifies a workflow that looks automatable. It runs frequently, it's mostly rule-based, it takes a lot of time. So they build an agent for it. The agent follows the documented process. The documented process is wrong.

It's wrong because the documentation was written two years ago, or because three people do the workflow differently, or because the real process includes workarounds and judgment calls that never made it into any document. The person who does the work knows the shortcuts. They know which clients get special handling. They know the spreadsheet that Dave maintains informally that's actually the system of record. None of that is written down.

The most common mistake in AI adoption is encoding existing inefficiency into a new system. Automating a broken process just produces broken outputs faster.

This is why the AGENTIC Framework puts Assess first in the AGENT Pipeline. Nothing enters the pipeline without it. And within Assess, the first job is not to document the workflow for automation. The first job is to fix it.


How AI workflow assessment actually works

Assess has three jobs. Map the workflow as it actually runs. Surface the hidden knowledge it depends on. Fix what's broken before you formalise it. Only then do you produce the specification.

Map the workflow as it actually runs

Not the documented version. Not the aspirational version. The real one, including every shortcut, workaround, and judgment call that lives in people's heads.

The AGENTIC Framework starts this with a conversation, not a template. The Conversational Audit Playbook provides the methodology: record a natural conversation with the person who does the work, let them talk through it, then use AI to extract the workflows from the transcript. People are much better at reacting to a draft than creating from scratch. The extraction is 70 to 80 percent accurate on first pass. The validation step gets it to 95 percent or higher.

Watch people do the work, not just hear about it. Live screen capture during conversations reveals steps that have become so automatic the person forgets to mention them. The tab switches, the quick checks, the "that's not really part of the process" things that absolutely are part of the process.

Surface the hidden knowledge

Every workflow depends on tacit knowledge: undocumented rules, judgment calls, informal shortcuts. These are the things a new starter would get wrong. They're also the things that will break an agent if they're not made explicit.

The hidden rules are where specifications break. Every workflow has steps that are documented and steps that live in someone's head. A person doing the work might apply approval thresholds, exception handling, or routing logic that nobody wrote down. If the specification doesn't capture those, the build will fail on real data. The gap between the documented process and the actual process is where Assess does its most important work.

Fix it before you formalise it

This is the step most assessment processes skip entirely. Before producing a specification, step back and look at the workflow's health. Where are the bottlenecks? Where are the redundancies? Is there an approval step that was added three years ago for a reason nobody remembers?

Score the workflow across clarity, consistency, efficiency, resilience, and documentation. If it scores poorly, fix it first. The optimisation happens here, not in Engineer. You are converging toward the best possible version of this workflow before you hand it to an agent.

Also assess the data. Are the data sources consistent, accessible, and structured enough for agent consumption? Data quality, format standardisation, and integration complexity are often the hardest practical barriers to automation. Catching these at Assess is cheaper than discovering them at Engineer.


From assessment to specification

Once the workflow is mapped, the hidden knowledge surfaced, and the process optimised, you translate it into an executable specification. The AGENTIC Specification Generator handles this: a prompt-driven agent that takes the assessment outputs and produces a two-layer specification.

Layer one is a human-readable playbook. A reference document anyone could follow, that the Workflow Owner signs off on. The playbook stays clean, concise, scannable. Layer two is a full structured specification with every input, output, decision gate, escalation path, failure mode, data requirement, and success criterion. The spec layer goes deep: full edge case libraries, worked examples of every decision path, exception catalogues with real data, conditional logic spelled out in full. Everything you'd tell a new starter over six months of sitting next to them goes in the spec. The agent doesn't skim. It uses all of it. The two layers intentionally diverge. Don't constrain the spec layer to what a human would want to read. Both layers must describe the same workflow, but readability is the playbook's job, not the spec's.

For high-risk workflows, the specification includes an additional layer: governance checkpoints. For each step with a governance implication, the spec captures what could go wrong, what control mitigates it, and what evidence proves the control operated. This layer is proportional to risk. A low-risk internal formatting workflow doesn't need it. A regulatory filing does.

The first specification will have gaps. That's normal. The refinement loop runs repeatedly: share the playbook with the Workflow Owner, walk through it, catch the missing rules and edge cases, iterate. First-pass specifications are typically 70 to 80 percent accurate. The loop gets them to 95 percent or higher.

A first-pass specification is typically 70 to 80 percent accurate. That's the point. The refinement loop is where accuracy happens, and it only works if you start with real conversations, not documentation.

Success criteria need to be specific and measurable. Not 'works well' but defined thresholds: accuracy rates, resolution times, compliance requirements. These become the acceptance criteria at Engineer and the monitoring thresholds at Nurture.


What the Assess stage produces

Here's the full flow from conversation to specification:

Conversation AI extraction Validation Tacit knowledge map Health assessment Optimisation Specification Refinement loop Validated spec

A workflow exits Assess with a validated, machine-readable specification that includes: the workflow mapped as it actually runs, tacit knowledge surfaced and documented, bottlenecks and redundancies addressed, organisational readiness assessed, success criteria defined, test cases designed, and governance checkpoints embedded.

That specification is what Greenlight scores, what Engineer builds from, what Nurture monitors against, and what Track resurfaces when capability changes. Every downstream stage depends on what Assess produces. Skip it, and you're building on sand.

The specification is not paperwork. It's infrastructure. Agents can read it, score it, and surface improvements when capability changes. Documentation that machines can parse becomes an asset that compounds.


What I've learned applying this

Two patterns show up every time.

First: people consistently underestimate how much hidden knowledge their workflows depend on. The finance manager with five approval rules thought there were two. The conservation team's data pipeline had an informal spreadsheet that was the actual system of record. In every case, the "simple" workflow turned out to have layers of undocumented complexity that would have broken any agent built without uncovering them.

Second: the optimisation step changes more than you'd expect. Removing a single redundant approval cut the cycle time of one workflow by 40 percent before any automation was involved. Unifying cost centre codes eliminated a category of monthly errors entirely. The workflow that enters the specification is often materially different from the one you started with. And that's the point.

Assess isn't fast. A thorough assessment of a complex workflow takes days, not hours. But every hour spent here saves multiples downstream. A specification built on a misunderstood workflow costs you at Engineer, again at Nurture when overrides pile up, and again at Track when you realise the whole thing needs rebuilding. Get Assess right and the rest of the pipeline flows.