Skip to main content Skip to footer
  • "com.cts.aem.core.models.NavigationItem@318f8cf4" Careers
  • "com.cts.aem.core.models.NavigationItem@dc7d30e" News
  • "com.cts.aem.core.models.NavigationItem@251fef2b" Events
  • "com.cts.aem.core.models.NavigationItem@64f9d024" Investors
Cognizant Blog
Most organisations rush to deploy AI and then wonder why it doesn't scale. The ones that get it right build the foundations first (e.g. governance, infrastructure, and clear ownership) before a single model goes to production.
 
The AI Rush Has a Hidden Cost

There is a pattern playing out across enterprises right now. A business unit sees a competitor launch a GenAI feature. An executive sponsors a proof of concept (PoC). The PoC gets positive feedback.

Eventually someone asks: "Can we productionise this?"

That's the moment organisations discover they don't have the strong foundations to answer “Yes!” for various reasons:

  • No clear ownership of the model.
  • No audit trail for its outputs.
  • No process for what happens when it drifts.
  • No way to scale the pattern to the next use case without starting from scratch.

The organisations that are scaling AI successfully in 2026 share one trait: they built a Target Operating Model (TOM) for AI before they had dozens of models in production — not after. This post lays out what that looks like, from governance foundations to the specific pillars that make scaling AI real and tangible.
 

The Five Layers of an AI TOM

Think of an AI operating model as a stack. Each layer must be solid before the one above it can function reliably. Skip a layer and you'll feel the gap (usually in production), at the worst possible moment.

1. Governance & Risk

Every organisation needs to ask fundamental questions when it comes to leveraging AI: Who owns AI decisions? How are use cases approved? What happens when a model underperforms or produces harmful output? The governance layer solves that.

This layer cuts across everything else — it's not a phase you complete once. It's a permanent, live capability.

2. AI Strategy & Use Case Portfolio

A structured way to decide what to build — and what not to. To make things simple, categorise AI initiatives in three buckets – AI for efficiency, AI for augmentation or AI for enterprise transformation. Prioritise on value versus feasibility.

The trap most organisations fall into is - over-investing in GenAI because it's visible, while neglecting the data quality work that actually determines model performance.

3. Data Infrastructure

A Lakehouse architecture is the current enterprise standard - a unified storage layer accessible to both analytics and AI workloads. Note here that GenAI adds a new first-class concern: vector stores for RAG pipelines.

These are no longer optional; they're as fundamental as your data warehouse.

4. MLOps & AI Engineering Platform

This is the operational backbone. It covers experiment tracking, model registries, CI/CD pipelines, drift monitoring, and feedback loops.

For GenAI, extend this to LLM operations (or simply LLMOps - the set of practices, tools and processes for developing, deploying, and maintaining Large Language Model applications in production): prompt versioning, evaluation frameworks, guardrails, and token cost management elements come into play here.

For agentic AI, consider adding stateful orchestration, observability and guardrails (including circuit breakers) and human escalation paths (Human-In-The-Loop).

5. People, Roles & Ways of Working

Technology choices are secondary to whether the organisation can sustain the model. The most mature pattern is a federated Centre-of-Excellence (CoE): a central team owns standards, platforms, and reusable components while domain-embedded squads own use case delivery.

Avoid going to full extremes such as fully centralised operations (bottlenecks) or fully decentralised (governance gaps) operations. A hybrid model for most organisations is the best approach.

The TOM you build for an organisation standing up their first AI use case is fundamentally different from the one you need when you have thirty models in production. Be explicit about where the organisation is on that maturity curve, that the TOM needs designing for.
 

Governance is not a Checkbox

Most organisations approach AI governance in a similar way to regulatory compliance like GDPR. The problem with this mindset is it produces checkbox governance that protects no one.

Effective AI governance is built around three structural decisions:

1. Separate accountability clearly

You need three distinct structures, not one.

  • An AI Ethics/Risk Board sets policy and owns escalation for high-stakes decisions.
  • An AI Centre of Excellence (CoE) sets technical standards and guardrails. 
  • Product/Domain teams own individual AI assets and their outcomes.

The common failure is letting the CoE make business risk decisions or worse, having no clear owner at all. A word of caution here: even a well-intentioned CoE can become a bottleneck if it never evolves. When every use case must queue for central resource, business units start shadow-building to bypass the wait and eventually, the Governance collapses. The signal to watch for is a growing backlog of AI requests alongside teams hiring their own "AI people" outside the CoE's visibility.

2. Tier your risk

Not all AI presents an equal risk to stakeholders. For example, a GenAI chatbot answering HR FAQs does not belong in the same risk class as an AI Agent executing financial transactions.

To visualise this, build a simple 2×2 matrix of impact severity × autonomy level. Your governance process, approval requirements and monitoring intensity should map directly to that classification.

3. Connect AI Governance to Data Governance

AI Governance and Data Governance cannot operate separately. Data lineage, access controls and quality SLAs are inputs to model reliability and audit readiness. If your data governance is weak, your AI governance has no substance. The EU AI Act, GDPR, and FCA guidance for financial services all assume a level of data traceability that most organisations haven't yet achieved.

Governance is not a layer at the top of the stack. Rather, it's a vertical capability that cuts across every plane of the operating model.
 

The Golden Path: Do One Thing First and Get It Right

One of the most common failure modes in enterprise AI programmes is trying to build the entire operating model before delivering any value. Organisations spend six to twelve months writing policies, hiring teams, procuring platforms, and running workshops — and emerge with a governance framework but no actual AI capability.

The alternative is the Steel Thread approach: identify one high-value, manageable-risk use case, build the minimum viable operating model to support it, deliver it, prove the value, and then expand on this with the pattern you've proven as your template. This will help industrialise the operating model and ways of working faster and will scale quicker.

Here's how to think about sequencing this:




How do AI Agents and Agentic AI impact the Operating Model?

Traditional AI/ML and GenAI have one thing in common: they're fundamentally stateless. You send a request; you receive a response. Agentic AI breaks that model entirely.

Agents are stateful, long running and capable of taking consequential actions such as querying databases, calling APIs, writing code, sending communications etc.

This changes what your operating model needs to handle:

  • Orchestration is now a compliance requirement 
    Every tool an agent can now call needs to be logged and attributable. "The model decided to do that" is not an acceptable audit response. Use Step Functions (AWS) or Durable Functions (Azure) to create durable, inspectable workflow records, not just inference logs.

  • Human escalation paths must be designed in advance
    Define, before deployment, the conditions under which an agent pauses and escalates to a human. Low-confidence decisions, high-value transactions, novel situations outside training distribution. This is a product design decision, not a model tuning decision.

  • Memory management is an engineering problem
    Short-term context, long-term persistent memory, and episodic recall all need to be architected explicitly. Agents that lack memory management either hallucinate context or balloon token costs. Neither is acceptable in production.
     
Three Questions That Unlock the Right Conversation

When we advise clients on their AI operating model design, the diagnostic conversation matters more than the framework.

These three questions consistently surface the real maturity gaps:

  • "Where is your AI risk sitting today, and who owns it?" If the answer is vague (e.g. - "the data science team" or "IT risk") then you've found your first gap. Ownership of AI risk needs to be explicit, named, and accountable. Not a committee. Not a policy document. A person.

  • "Show me a model you have in production and tell me how you know it's performing." This question exposes the operational maturity almost instantly (more specifically – MLOps/LLMOps). Most organisations can show you a model. Very few can show you a monitoring dashboard, a drift alert that fired last month or a retraining process that responded to it.

  • "What does your AI use case pipeline look like in 18 months?" The TOM you design for two use cases is not the TOM you need for twenty. Understanding your enterprise ambition determines whether you're designing a starting model or a scaling model as those are genuinely different architectures.

To summarise, AI delivers compounding returns - but only when the foundations compound too. A well-designed operating model means each new use case is cheaper to deliver, safer to govern, and easier to scale than the last. Without it, each new use case is effectively a new programme.

The organisations winning with AI in 2026 aren't necessarily the ones with the most sophisticated models. They're the ones who made the less glamorous investment in governance structures, data contracts, MLOps pipelines, and clear ownership and who had the discipline to build those things before they felt urgent.


Cognizant UK & Ireland
Author Image

Our experts are contributing exciting insights about what's going on within technology and innovation.





In focus

Latest blog posts

More blog posts