Here's what happens in most AI agent deployments: someone builds the agent, tests it informally ("looks good to me"), and ships it to production. The team is told it works. There's no structured evidence that it does. The first time anyone sees real performance data is after the agent is already running live workflows, handling real data, and making real decisions.

If the agent underperforms, trust collapses immediately. Not just in that agent, but in the organisation's entire AI effort. One bad deployment can set an adoption programme back months.

The Engineer stage in the AGENTIC Framework exists because shipping an agent without evidence is a trust problem disguised as a technical one. Engineer takes the validated specification from Assess and the scored collaboration model from Greenlight, builds in sandbox, proves through parallel-run testing, and deploys only when the evidence says it's ready. Every workflow passes through this. It is non-negotiable.


Why most AI agents go live without proof

The pressure to ship is real. Leadership wants results. The team that built the agent is confident. The demo looked great. So it goes live, and the first real test is production.

The problem is that demos don't test edge cases. They don't test the fourth data source that has inconsistent column headers. They don't test what happens when the field team changes their spreadsheet format seasonally. They don't test the approval rule that lives in one person's head and never made it into the specification. These are the things that break agents in production and they only surface when you test against real data, in real conditions, over real time.

Parallel-run testing solves this. The agent runs alongside the human process. Outputs are compared side by side. Discrepancies are logged and investigated. You build trust through evidence: quantified performance data that proves the agent can handle the workflow before anyone is asked to rely on it.

The parallel-run generates trust through evidence. Skip it and you're asking the organisation to trust on faith.


How AI agent builds actually work

Engineer has a clear sequence: build in sandbox, test against real data, evaluate against the success criteria defined at Assess, prove through parallel-run, then deploy. The collaboration model from Greenlight is the engineering blueprint. It defines exactly which steps the agent handles, at which autonomy level, with which human checkpoints.

Build in sandbox, not production

The sandbox is an isolated environment that mirrors production. It exists so the build can fail safely. Start with the AI-run steps from the collaboration model, the ones that are fully self-contained and easiest to prove. Get those working first. Then layer in the AI-led steps where human verification is needed. Then configure the handoff points between human-run and agent-run steps. Each layer is proven before the next is added.

A sandbox prototype is rarely complete without the person who does the work. They bring the judgment, the system access, and the edge-case knowledge that turns a proof-of-concept into something that actually connects up.

The Workflow Owner is involved throughout the build, not waiting for a finished product. At the finance team build I worked on, the sandbox prototype connected to three of four data systems on the first attempt. The fourth, an informal shared spreadsheet the field team maintained, had inconsistent column headers that broke the import. The finance manager identified the issue in minutes: the field team changes column names seasonally. A normalisation rule was added to the specification, and the data pipeline worked across all four sources. That's the kind of thing no specification catches on paper. You need the person in the room.

Run the build against real data. Then run the test cases defined at Assess: a set of inputs and expected outputs covering the full range of the workflow, including the happy path, edge cases, and known failure modes. Measure performance against the success criteria. Document what passes and what fails. If something fails, diagnose whether it's a build issue, a specification gap, or a success criterion that needs revisiting. Fix it and retest. This loop continues until the build meets every criterion.

AI agent parallel-run testing is non-negotiable

Once the build passes its test cases in sandbox, the parallel-run begins. Two kinds of testing matter. First: does the design work? Walk through the workflow end-to-end with real data, verify each step produces what it should. The parallel-run handles this. The agent processes real workflows alongside the human. Outputs are compared. Every discrepancy is logged and categorised: was it an agent error, a specification gap, or a case where the agent was actually right and the human process had an inefficiency?

The parallel-run should cover at least one full cycle of the workflow. The agent runs alongside the human process, and at the end you compare outputs step by step. Where they match, confidence is high. Where they diverge, you investigate: is it a configuration issue, a missing rule, or a genuine limitation? The data from the parallel-run is what gives you the evidence to ship. Whoever built the workflow shouldn't be the only person evaluating it. Fresh eyes catch different things. Bring in someone who wasn't involved in the build to review the parallel-run results.

That data is what moves people. Not a demo. Not a promise. Quantified performance over a real cycle, compared against how the work was done before. The parallel-run also generates the performance baseline that Nurture will monitor against once the workflow is live.

Not everything is a bespoke agent

The word "build" implies custom development, but that's often not what Engineer produces. The build might be a Notion database with AI-populated columns. A Copilot workflow in Microsoft 365. A Slack automation. A pre-configured template in an existing tool. A combination of several. The framework doesn't prescribe what technology you use. What matters is that the specification is implemented, the collaboration model is honoured, and the parallel-run proves it works.

Sometimes the build is a Notion database with AI columns, a Copilot workflow, or a pre-configured prompt. The specification and collaboration model apply regardless of how the technology is assembled.

This is important because it keeps Engineer accessible. You don't need a team of AI engineers. You need someone who understands the tools the organisation already uses and can configure them to match the specification. Templates, database structures, and pre-configured prompts aren't shortcuts. They're legitimate solutions when they deliver what the specification requires.


What the Engineer stage produces

Here's the full flow from specification to deployment:

Specification + collaboration model Sandbox build Real-data testing Test cases evaluated Parallel-run Production deployment Nurture handoff

A workflow exits Engineer with sandbox test results, test case logs, evaluation evidence, parallel-run performance data, a proven collaboration model, production configuration documentation, and a deployment record. Everything goes into the AGENTIC Vault. Everything is structured so that Nurture can monitor against the baselines Engineer established.

Rollback triggers are defined before deployment, not after something breaks. If accuracy drops below the threshold defined at Assess for two consecutive cycles, the step reverts to human-run pending investigation. If escalation routing fails on any compliance-flagged item, the workflow pauses and alerts the Workflow Owner. These aren't afterthoughts. They're part of the build.

Engineer also feeds back into the pipeline. Every specification gap found during sandbox testing gets fed back into the Assess artefacts. Every collaboration model adjustment gets fed back into Greenlight. The specification and collaboration model co-evolve through the build. What exits Engineer is often materially better than what entered it.


What I've learned building this way

Two things stand out.

First: the sandbox phase catches problems that no amount of specification review will find. System connections that weren't documented. Data formats that change seasonally. Performance characteristics that only show up under real conditions. Every build I've worked on has discovered at least one significant specification gap during sandbox testing. That's the point. You want those gaps to surface in sandbox, not in production.

Second: the parallel-run changes the adoption conversation entirely. Before the parallel-run, people are nervous. They've been told the agent works. They haven't seen it. After the parallel-run, they have data: accuracy percentages, turnaround times, a log of every discrepancy and how it was resolved. The conversation shifts from "do we trust this?" to "look at what it does." That shift is what makes deployment smooth. Not the technology. The evidence.

Engineer is the stage where the framework's documentation-first approach pays off most visibly. The specification from Assess becomes the build blueprint. The collaboration model from Greenlight becomes the engineering plan. The success criteria become the test cases. Nothing is invented at Engineer. Everything is executed from artefacts that already exist. The build is a translation, not a creation. And that's what makes it repeatable.