/

February 9, 2026

Agentic AI needs guardrails before it gets autonomy

Agentic AI needs guardrails before it gets autonomy

Recent discussions with US and Canadian senior decision-makers indicated that agentic AI is moving from concept to real-world experimentation across industries, including healthcare, financial services, hospitality, and insurance. The appeal is obvious: systems that can make independent decisions with minimal human intervention can remove friction from workflows that have historically been slow, manual, and costly.

The same discussions were equally clear on the risk. The more autonomy you give an agent, the more you must invest in trust, governance, and human oversight. In practice, agentic AI is not a technology decision. It is an operating model decision.

This article translates what emerged in those discussions into a production-minded blueprint: where agentic AI can help, why it fails without guardrails, and what “safe autonomy” should look like in 2026.

What agentic AI is actually changing

Decision-makers described a shift from human-led workflows to AI-driven workflows, especially for backend processes, data entry tasks, and scalability demands. The difference with agentic AI is not only that it can generate content or insights. It can choose actions.

That shift changes the risk profile immediately. When a system can initiate actions, errors become operational incidents, not analytics disagreements.

Several practical initiatives were discussed that illustrate where organisations are aiming first:

  • Voice AI designed to handle 10 to 20 percent of customer calls
  • Intelligent document processing aimed at automating 70 to 80 percent of data ingestion from multiple systems
  • Agentic systems intended to manage disputes more efficiently

These are high-value targets, and they sit inside workflows that touch customers, regulators, financial outcomes, and brand trust. That is why guardrails are not optional.

Why “minimal human intervention” is not the same as “no human oversight”

A consistent theme was that agentic AI adoption is still in its infancy in many environments, and leaders are reluctant to expand autonomy without high accuracy. That stance is rational.

In parallel, the discussions emphasised the need for human supervision at critical control points. The implication is simple:

  • Autonomy can exist inside a workflow
  • Accountability cannot be outsourced to a workflow

This is where many programmes stumble. Teams interpret “autonomous” as “hands-off.” The leaders in these discussions framed it differently: the goal is to reduce routine effort, while increasing oversight where risk and impact are highest.

The three prerequisites for safe autonomy

Across the conversations, three prerequisites repeatedly surfaced as non-negotiable.

1) Process mapping before automation

Participants agreed on the importance of thorough process mapping before agentic AI is given authority. If you cannot describe the workflow, you cannot safely automate decision-making inside it.

Process mapping is more than documentation. It clarifies:

  • Which decisions are reversible and which are not
  • Where human judgement is required
  • Which data signals matter, and what happens when those signals are missing
  • Where bias can enter the system
  • What the escalation path should be when confidence is low

Agentic AI is only as safe as the workflow it operates inside.

2) Trust in the model and the data

Trust came up repeatedly, including the need for appropriate human oversight to ensure data accuracy, and the need to determine acceptable error thresholds before production deployment.

Trust also intersects with data governance and quality. Leaders discussed challenges around data centralisation and integration, and the operational concern of data segregation and recovery in disaster scenarios.

The message was consistent: if your data foundation is fragmented, your agents will behave inconsistently. Worse, they may behave confidently while being wrong.

3) Human oversight designed into the system

Participants repeatedly expressed concern about AI making autonomous decisions without human oversight, especially in regulated contexts where ethical decision-making and compliance cannot be delegated to a model.

This is not a philosophical point. It is the practical difference between a helpful system and a liability.

A practical autonomy model for 2026

A useful way to structure agentic AI deployment is to define autonomy levels based on risk and reversibility.

Graph: autonomy level vs guardrail intensity (higher bar means stronger guardrails required)

  • Suggest only (agent recommends, human decides): ██
  • Execute low-risk actions (agent acts, human can easily undo): ███
  • Execute medium-risk actions (agent acts, limited reversibility): █████
  • Execute high-risk actions (agent acts, customer or financial impact): ██████

The discussions repeatedly implied that most organisations should start at “suggest” or “low-risk execute,” then earn the right to expand autonomy based on measured performance and governance maturity.

Where agentic AI is being applied first

Leaders discussed examples across multiple environments. The common thread was that early agentic AI use is gravitating toward operational workflows with measurable throughput and clear failure detection.

Healthcare operations

One example discussed was the automation of inventory management for life-saving equipment. This is a strong use case because the workflow is defined, the operational impact is measurable, and the cost of human delay can be high.

The risk is also obvious: incorrect decisions can create safety exposure. That makes guardrails and oversight essential.

Financial services and insurance operations

Participants discussed the move from predictive to generative approaches, and the exploration of agentic AI for backend process automation. There was also mention of governance-led proof-of-concepts for workflows such as underwriting and claims tooling.

These are areas where the upside is significant, but the tolerance for error is low, and the compliance footprint is large.

Hospitality and customer interaction

Agentic AI was discussed as a way to enhance guest interactions. Customer-facing workflows are where trust is easiest to lose, so these use cases require especially clear escalation and human handoff design.

Dispute management and document-heavy workflows

Agentic systems for managing disputes, and intelligent document processing to automate 70 to 80 percent of ingestion, are compelling because they reduce manual backlog and improve cycle times.

They also create a new measurement requirement: you must track where automation is failing, not only where it is succeeding.

The four ways agentic AI fails without guardrails

The discussions did not frame failures as “AI is bad.” They framed failures as “AI is unmanaged.”

Failure mode 1: Accuracy is assumed, not proven

Leaders emphasised the need for high accuracy before widespread deployment and the importance of setting acceptable error thresholds before production.

Without thresholds, teams cannot answer basic questions such as:

  • How wrong is too wrong?
  • What happens when the agent is uncertain?
  • When do we require human confirmation?
  • What do we do when outputs drift over time?

Failure mode 2: Bias is ignored until it becomes a problem

Bias was repeatedly raised as a concern in agentic AI adoption. In the responsible AI discussions, fairness and accountability were emphasised, along with the need for diversity in models, stress testing, and human oversight.

The practical point is that bias is easier to catch early, when autonomy is limited, than later, when the agent is embedded in core workflows.

Failure mode 3: Governance exists, but it is not operational

Responsible AI was discussed in practical terms: data governance, human oversight, audit trails, and risk assessment. Governance was also described as requiring acceptable use guidelines, training, and evaluation of use cases.

Where governance fails is when it stays in policy documents. Where it works is when it becomes workflow steps, training expectations, and enforceable access controls.

A particularly useful principle raised was that organisations should evaluate low-risk tools quickly, while requiring pilots and buy-in for higher-risk deployments. That is a governance model that matches how autonomy should expand.

Failure mode 4: Security is treated as separate from agent behaviour

Multiple discussions on cybersecurity highlighted that AI can increase threats, and that automation must be balanced with human oversight. Participants also expressed concern about widespread implementation of AI agents without proper security considerations, comparing it to historical mistakes where databases were exposed to the internet.

The implication is important: agents expand the attack surface because they interact with tools, data, and workflows.

If you do not design security and access control into agent behaviour, you will not control the blast radius of misuse, prompt injection, data leakage, or unauthorised actions.

The guardrails table: what to implement before autonomy expands

The table below consolidates the guardrail categories repeatedly referenced across the discussions and turns them into a practical pre-production checklist.

Guardrail categoryWhat it protectsWhat leaders emphasised in practiceA practical implementation approachA leading indicator to monitor
Process mapping and workflow designPrevents agents acting in undefined or unsafe stepsThorough process mapping, control points, and workflow shift designMap decision points, define reversibility, define escalation and handoffPercentage of workflows mapped before pilot expansion
Data governance and qualityPrevents confident wrong actionsData governance and quality issues, data centralisation and integration challengesStart with a small set of trusted fields, enforce validation and ownershipData exception rate and validation pass rate
Trust framework and thresholdsPrevents uncontrolled errorDetermine acceptable error thresholds before productionDefine confidence thresholds and “stop” conditionsPercentage of actions executed within thresholds
Human oversight and accountabilityPrevents ethical and compliance failuresHuman supervision at critical control points, concerns about autonomous decisionsRequire human confirmation for high-impact actionsHuman review coverage on high-impact decisions
Bias and fairness controlsPrevents systematic harm and reputational damageAddressing biases, fairness and accountabilityStress test models, include bias checks in evaluationBias exception flags per workflow
Audit trails and traceabilityEnables investigation and compliance defenceAudit trails were emphasised as part of responsible AILog actions, inputs, approvals, and changesAudit completeness for each agent action
Training and acceptable usePrevents shadow usage and inconsistent behaviourTraining, acceptable use guidelines, continued educationCreate role-based guidance and approval pathwaysTraining completion and policy adherence checks
Security and access controlsPrevents data leakage and unauthorised actionsAI governance committees, access control concerns, shadow IT riskImplement least privilege, tool access restrictions, and monitoringUnauthorised access attempts and anomalies
Policy refresh cadenceKeeps governance relevant as tooling changesUpdating AI policy every 6 monthsReview policy twice per year with cross-functional ownersTime since last policy update

The security reality agents force you to confront

Several cybersecurity discussions described a rapidly shifting threat landscape.

One striking example referenced a new worm that infected up to 100,000 code repositories, raising concerns that developers could unknowingly download and run infected packages through normal processes. Participants also highlighted that old attack vectors keep reappearing in new forms, and that vulnerabilities can emerge in popular tools even when teams believe they have secured their AI stack.

Agentic systems amplify this risk because they are designed to do work on behalf of humans. If an agent has access to sensitive systems or tooling, it can inadvertently accelerate compromise.

Leaders discussed practical governance responses, including forming AI governance committees and increasing scrutiny on identity management and access controls in organisations where AI tools are spreading rapidly.

A particularly memorable access-control example involved demonstrating that an AI assistant could access password-protected information, triggering a broader discussion about governance and access controls. The point is not the tool. The point is what it revealed: AI can traverse permissions and surface information in ways that standard UX patterns do not anticipate.

Responsible autonomy requires responsible ownership

A question raised in the governance discussions was: who defines ethical data practices?

A clear stance emerged that the Chief Risk Officer role is often the accountable owner, working with risk, security, and legal teams to develop guidelines and policies. There was also discussion of working with affected communities and the role of lawmakers in shaping ethical frameworks.

This matters because agentic AI compresses decision-making cycles. If ownership is unclear, decisions stall, or worse, they are made informally without governance.

A pragmatic operating model surfaced:

  • Accelerate low-risk tools and use cases quickly
  • Require pilots, buy-in, and stronger oversight for higher-risk use cases
  • Slow down where long-term risk outweighs short-term speed

That mindset is the foundation of safe autonomy.

How to launch agentic AI without losing control

Based on what emerged in the discussions, a controlled rollout should follow a staged approach.

Stage 1: Suggest mode inside one workflow

Start with an agent that recommends actions, but does not execute.

Focus areas raised in the discussions that suit this stage include:

  • Back-office workflow assistance
  • Data entry support and triage
  • Dispute case categorisation
  • Document routing and prioritisation

The advantage of suggest mode is that it produces learning without operational risk.

Stage 2: Execute only reversible, low-risk actions

Once thresholds and trust are proven, allow execution where mistakes are easy to undo.

Examples aligned to the discussion themes include:

  • Automating parts of ingestion workflows where exceptions can be routed to humans
  • Low-impact customer service tasks with clear escalation paths

This is where you start to see measurable throughput gains while maintaining safe control.

Stage 3: Expand autonomy only after governance and auditability are proven

Before expanding into underwriting, claims, or sensitive customer interactions, ensure:

  • Audit trails are complete
  • Human oversight coverage is measurable
  • Data quality ownership is clear
  • Security and access controls are enforced
  • Bias and fairness checks are operational

Leaders repeatedly implied that autonomy must be earned, not granted.

The operating model shift leaders should plan for

A key theme was that business units increasingly own data while IT manages infrastructure, and that business units must validate and clean data to maintain quality. This aligns with the reality of agentic AI deployment.

IT can enable and increasingly accelerate AI adoption, but autonomy cannot be safely scaled unless business owners also own:

  • Data validation
  • Workflow definition
  • Outcome measurement
  • Exception handling
  • Accountability for decisions made in the workflow

Participants also emphasised the need to incorporate change management from the outset, providing training and upskilling to address fears related to AI, and measuring success through business outcomes such as improved revenue, decreased costs, and increased efficiency.

Agentic AI will intensify all of those needs.

Closing thought

Recent discussions with US and Canadian senior decision-makers indicated that agentic AI is most valuable where it removes repetitive operational load and improves scalability. Those same discussions were clear that autonomy without guardrails is an avoidable mistake.

Safe autonomy requires:

  • Process mapping
  • Trust thresholds
  • Strong governance with auditability
  • Human oversight at critical control points
  • Security and access controls that match an expanded attack surface
  • Training and policy refresh cadence that keep up with change