Here's what happens after most organisations deploy their first AI agent: nothing. The agent runs. Nobody watches it. Nobody asks whether the technology it was built on is still the best option. Nobody checks whether workflows that were rejected six months ago are now feasible. Nobody has a plan for what to do when the agent should stop running.

The first deployment goes well. Maybe the second does too. Then the programme stalls. Not because the technology failed, but because there's no mechanism to decide what comes next, no way to spot when the landscape has shifted, and no governed path to retiring what no longer works. The pipeline runs once and stops.

The Track stage in the AGENTIC Framework exists to solve this. Track is the final stage of the AGENT Pipeline and the reason the pipeline compounds. Where Nurture keeps live workflows healthy, Track looks outward: it watches the AI capability frontier, resurfaces workflows that are newly ready, monitors cost and performance across the portfolio, and manages the governed path to decommissioning workflows that no longer serve the organisation. Without Track, you have a deployment. With Track, you have a learning system.


Why AI agent portfolios stall

Most AI adoption programmes are built around a single question: what should we automate? They answer it once, build the thing, and move on. The problem is that AI capability changes constantly. A workflow that was scored "not yet" at Greenlight because the technology couldn't handle ambiguous document parsing might be perfectly feasible six months later when a new model ships. But if nobody is watching the frontier and mapping it against the scored backlog, that workflow stays parked forever.

The same is true in reverse. A workflow that was cost-effective when the team was small might become uneconomical when the organisation restructures. An agent built on a tool's AI feature might need retiring when that tool sunsets the capability. These are predictable failure modes, but most organisations only discover them after something breaks.

The biggest risk after a successful AI deployment isn't technical failure. It's the programme stalling because nobody built the mechanism to decide what comes next.

Track turns a one-shot deployment into a continuous cycle. It keeps the AGENT Prioritisation Matrix alive, feeds resurfaced workflows back through Greenlight, and ensures decommissioning gets the same governance rigour as greenlighting. The organisation never has to manually ask "what should we automate next?" or "when should we stop?" Track surfaces the candidates and the retirement triggers. Humans make the calls.


How Track watches the frontier

Track monitors two things: the broader AI landscape and the organisation's own software stack. The broader landscape is models, research, and tooling. The organisation's stack is where the most accessible shifts happen: Notion ships a new AI agent feature, Copilot adds a capability, Slack rolls out workflow automation, a CRM embeds intelligence it didn't have last quarter. These are often the changes that matter most because they don't require building anything new. They require recognising that something you already use can now do something it couldn't before.

When new capability surfaces, Track maps it against every workflow in the AGENT Prioritisation Matrix. The matrix holds every workflow that has ever been scored, whether it was greenlighted, deferred, marked human-run, or retired. Which workflows had steps that were previously marked "not yet" because the technology couldn't handle them? Which collaboration models could shift because a step that was human-led might now be AI-led? Track identifies the change, the assessment agent rescores the relevant workflows, and the matrix reshuffles.

Resurfacing is not re-greenlighting

The word is deliberate. Track resurfaces workflows. It flags them with evidence: what changed, what the updated cost picture looks like, why the readiness status is different. But the decision to commit still goes through Greenlight. The capability landscape has changed, so the scores change. The collaboration model may need redesigning. The cost-benefit case is different. Greenlight decides whether to build. Track provides the intelligence. The pipeline's decision gates stay intact.

Track also monitors cost and ROI across the portfolio. Cost per run, cumulative cost, ROI trends. A workflow that was cost-effective at one scale might become uneconomical if volume drops. Equally, a workflow that's expanding in scope becomes more valuable. This is a live dimension that shifts as the organisation changes. Exit criteria for each pipeline stage reference specific Vault artefacts that must exist. A workflow exits Assess when the specification, success criteria document, test case library, and governance checkpoint map are all validated. An agent could scan the Vault, check whether required artefacts exist for a workflow at its current stage, and flag what's missing. The gate becomes machine-verifiable.


AI agent decommission strategy

This is the part most frameworks skip entirely. Not everything that gets automated should stay automated forever. Performance drops below threshold. The tool an agent was built on gets deprecated. A governance model is invalidated by new regulations. A team restructures and the workflow owner is no longer available. The business need disappears. These are all legitimate triggers for retiring an AI workflow, and they need a governed process.

A workflow doesn't just get quietly turned off. The decommission path requires the same governance rigour as the greenlight path.

The AGENTIC Framework treats decommissioning as a first-class governance event. When Track identifies a decommission trigger, the process is structured: the trigger is documented, governance reviews and signs off on the retirement, the AI Adoption stream plans the transition for affected teams, stakeholders are notified, and the workflow moves to Retired status in the matrix. All specifications, build files, and performance data are archived in the AGENTIC Vault. Nothing is lost. The knowledge compounds even when the workflow doesn't.

At one of the organisations I've worked with, a purchasing approval agent faced this exact situation. The accounting software it was built on announced it was sunsetting its AI capability in six months. Track identified this early, before anything broke. Governance reviewed the retirement plan. The AI Adoption stream prepared the team: they went back to a faster manual process temporarily while three replacement solutions were evaluated. No surprises. No abandonment. The team knew what was happening and why.

That's what a governed decommission looks like. Compare it to the alternative: the tool quietly degrades, the agent starts producing errors, trust collapses, and the team finds out when something fails in production. One approach preserves trust. The other destroys it.


The compounding advantage

Every workflow that passes through the AGENT Pipeline produces reusable artefacts: specifications, collaboration models, performance baselines, tested modules. These live in the AGENTIC Vault that Nurture maintains. The second workflow is faster than the first. The tenth is faster than the fifth. Track is the stage that makes this compounding visible and actionable.

Here's the full cycle:

Frontier shift detected Matrix mapping Workflows resurfaced Greenlight rescoring Next build committed Pipeline compounds

Track also identifies expansion opportunities beyond the current matrix. Adjacent workflows that share structure with proven ones are candidates. Reusable modules from the Vault that could unlock new parent workflows get flagged. Teams that have built trust through successful deployments and are ready for more get surfaced. But expansion is always gated by governance capacity. The pipeline cannot outpace what governance can cover.

Track connects what the organisation has learned to what the organisation should do next. That's what turns a deployment programme into a learning system.


What I've learned about the stage nobody builds

Two things stand out from building Track into real deployments.

First: the most valuable frontier shifts are usually the least dramatic ones. It's rarely a headline-making model release that changes the picture. It's the update to a tool the organisation already uses. Notion adds an AI feature that does what you were going to build a custom agent for. Copilot gains a capability that makes a deferred workflow suddenly feasible. These incremental shifts inside the existing software stack are where the fastest wins come from, and they're the easiest to miss if nobody is watching.

Second: decommissioning is the part that builds the most organisational trust. When a team sees that the framework has a governed process for retiring what's not working, they trust the framework more for what is working. It signals that the organisation isn't just chasing automation for its own sake. It's managing a portfolio with the same rigour coming out as going in. That signal matters more than most people expect.

Track is the stage that separates frameworks that run once from frameworks that compound. It's the reason the AGENT Pipeline is a cycle, not a sequence. The pipeline's first pass ships a workflow through Assess, Greenlight, Engineer, and Nurture. Track ensures the second pass is smarter, faster, and better informed than the first. Every cycle makes the next one easier. That's the compounding advantage, and it only works if someone builds the stage nobody thinks to build.