Trust by design: How much can you really trust your AI agent

A person typing on a laptop and using a tablet. Only their upper torso, arms and hands are visible. Text superimposed on the image shows AI
(Image credit: Getty Images)

When an AI system makes a consequential decision that your organization cannot fully explain, who is accountable for it?

It is a question that is becoming harder to avoid as systems that once waited for instructions begin to act autonomously, initiating tasks, making decisions, and adapting as they go.

For British businesses, this creates both a compliance risk and a strategic one, especially given the UK government’s clear ambition to accelerate the development of AI tools at pace with its £500 million Sovereign AI venture fund launching this April.

Article continues below
Ivana Bartoletti

Global Chief Privacy and AI Governance Officer at Wipro.

Consider a financial services firm encouraged to adopt an agentic AI to support credit decisioning, or a healthcare provider deploying a partner startup’s clinical triage assistant.

In both cases, the agent may be drawing on sensitive personal data, acting without direct human instruction, and shaping outcomes that carry real consequences.

The risk is made more pressing by something that rarely features in governance discussions: AI systems are becoming measurably more persuasive, particularly when they have access to personal context about their users.

Research shows that when AI knows something about who it is talking to, its persuasive capability grows more refined over time. In agentic systems with persistent memory, it compounds.

When users cannot tell why an agent responds as it does, or whether it is optimizing for their interests, trust slowly turns into dependency.

From compliance to trust by design

Most discussions of AI governance still revolve around harm prevention and regulatory compliance. These issues matter and always will. But preventing harm is not the same as shaping impact. In the age of autonomy, responsibility cannot be defined only by what does not go wrong. It must also account for the futures we are actively creating.

As AI agents move from tools to interlocutors, the core challenge becomes behavioral: how do we ensure these systems can actually be trusted? Trust by design means embedding that answer into the architecture of an agentic system from inception, not adding it on after deployment.

For organizations, it also represents a reframe: trust is not a barrier to adoption but a foundation for better outcomes, and increasingly a genuine competitive differentiator.

Earning rather than engineering that trust requires two distinct layers of design thinking: structural and psychological.

The trust stack

At the structural level, meaningful design means building a layered approach to autonomy. To trust what an AI system does, organizations need to understand what it knows, what it is allowed to do, and what it actually did.

That means starting with well-governed, traceable data, adding clear rules that reflect values and limits, and ensuring transparent decision records that allow actions to be questioned and learned from.

In practice, this means:

Legible reasoning paths: the agent should be able to explain how and why it reached an output, not as full technical disclosure but as meaningful traceability.

Bounded agency: clear limits on what the agent can do, decide or recommend, with no silent escalation of autonomy.

Goal transparency: the agent's objectives must be explicit. Users should know whether it is optimizing for accuracy, safety, efficiency, engagement or commercial outcomes.

Contestability and override: humans must be able to challenge, correct or disengage from the agent easily. Frictionless exit is a trust requirement.

Governance by design: logging, auditability and oversight mechanisms must be embedded from the start, not added later.

Before autonomy scales, there is an opportunity to slow down and observe. How does an agent behave once it is learning in the wild? What patterns does it start to favor? Do users defer more? Override less? Trust faster than they should?

Taking time to explore how those shifts play out is how organizations avoid sleepwalking into behaviors they never meant to normalize.

The psychological layer

People need to feel they still have agency, to understand when AI is acting and why, and to know how to intervene. Systems that are technically compliant but experientially opaque quickly erode trust. That demands deliberate design choices.

The agent should avoid anthropomorphic cues that suggest empathy or authority beyond its actual capacity, because emotional tone should not imply moral understanding.

It should signal uncertainty and confidence levels openly, because saying "I don't know" is a trust-building feature, not a limitation. It must not reinforce beliefs uncritically, mirror emotions to deepen attachment, or optimize for dependency. Trust built through such emotional mirroring is fragile.

The alternative is cognitive resonance: the quality of a system that behaves in ways users can intuitively understand, anticipate and critically interrogate.

This kind of trust holds up under scrutiny, because predictable, principled behavior builds more durable trust than adaptive influence. Cognitively resonant AI agents treat users as reasoning subjects, not behavioral targets.

A question worth sitting with

For any British business navigating both the opportunity and the scrutiny that comes with the UK government's AI ambitions, the reframe is significant.

The question leaders need to ask is not just "Is our AI responsible?" but what behaviors will this system normalize, what will it reward, and what will it quietly discourage?

The real test of responsible autonomy will not be the risks we avoided. It will be the futures we deliberately brought into being.

We've rated the best IT automation software.

This article was produced as part of TechRadar Pro Perspectives, our channel to feature the best and brightest minds in the technology industry today.

The views expressed here are those of the author and are not necessarily those of TechRadarPro or Future plc. If you are interested in contributing find out more here: https://www.techradar.com/pro/perspectives-how-to-submit

TOPICS

Global Chief Privacy and AI Governance Officer at Wipro.

You must confirm your public display name before commenting

Please logout and then login again, you will then be prompted to enter your display name.