Explainable AI: 4 key questions to check it hasn't become a compliance chore

Compliance is not the end. Explainability is the lever that turns models into outcomes, and a habit that keeps AI investments from drifting towards the drain.

Explainable AI: 4 key questions to check it hasn't become a compliance chore

The quickest way to make AI useful is to make it understandable. Not to everyone, all at once, but to the people who will live with its decisions, the product teams who will improve it, and the leaders who must defend it when the phone rings. Explainability is not icing for compliance. It is the lever that turns models into outcomes, and a habit that keeps AI investments from drifting towards the drain.

Let's start with a simple example. A loan is declined, and the customer asks why. If the answer cannot be explained in words the customer can accept, value and loyalty start to evaporate. Complaints follow, colleagues lose confidence, the models get quietly sidelined, and the promised productivity never lands. When the reasons are clear enough to challenge and change, something very different happens. Teams improve their data, product owners refine their guardrails, and customers accept decisions even when they are not happy with them.

Explainability does not mean publishing equations or every weight used to train a model. It means being able to show, with evidence, how an input became an output, what trade-offs were involved, and what a reasonable person can do if the decision feels wrong. That level of clarity reduces fear, invites better questions, and speeds up improvement.

There is another benefit that is often missed. Clear explanations expose weak incentives. When a model is optimised for the wrong signal, for example clicks instead of customer satisfaction, a good explanation will surface the mismatch. That creates a chance to correct purpose rather than fight symptoms. It also reframes bias as a design decision. Shortcuts, skewed samples and proxy targets do not remain neutral errors, they become choices about who carries the cost. Explainability makes those choices visible and open to challenge, so teams can rebalance incentives, improve data, and set guardrails that reflect the outcomes they actually want. It also prompts a more honest question, what kind of intelligence are we building, and how will it learn from the world we have. If the world is noisy or unfair, an accountable system will learn with scepticism, not blind imitation. Explanations keep that scepticism alive by showing what the model paid attention to and what it ignored.

To keep this practical, leaders should focus on four key conversations and make them routine.

  1. How does this model behave, and where does it fail? Teams should point to examples of correct and incorrect behaviour and describe patterns that are likely to break. Include plain language limits. For example, the model struggles with very short notes, confuses negation in sentiment, misreads dates across time zones, or loses context in multi turn chats.
  2. What data did we use, and what could be missing? A short data sheet and lineage note should explain sources, exclusions, and known gaps. Call out recency, geographic skew, device bias, and missing negative examples. For instance, training mostly on weekday traffic can degrade weekend performance, image sets from bright rooms fail on dim photos, and surveys from early adopters may not represent mainstream users. If downstream groups cannot see this, mistrust will grow, and problems will be hard to repair.
  3. What is the human's role? People need to know when they may override, how to appeal, and what happens next. Provide concrete guardrails that a non-specialist can remember. For example, a support agent can approve a goodwill refund up to £100, a moderator can restore obviously satirical content with one click, and a supervisor will review all overrides weekly for learning. If the human in the loop is symbolic, accountability will still be missing.
  4. What is monitored when the model is live? Drift, bias, abuse, and security should sit on a small scoreboard that a non-specialist can read. Useful checks include acceptance and complaint rates by segment, false positive and false negative rates on a small, labelled sample, alerts when top features or prompts change suddenly, throttles when traffic spikes from a single IP range, and a monthly fairness review across relevant groups. Surprises will still occur, but they will be found sooner, and fixes will be faster.

These conversations become easier if we reduce complexity at the design stage. Favour features that make sense to a domain expert. Keep models as simple as the problem allows. Use interpretable components where practical and add post hoc explanations where the domain truly needs depth. Protect privacy by minimising sensitive attributes and by testing fairness against them rather than training directly on them. None of this eliminates performance, it creates a surface where improvement is possible without fear.

A few actionable moves help this take root:

  • Writing a 1-page decision card for any AI that touches customers or staff: purpose, inputs, limits, override, appeal, and monitoring
  • Running reasonableness tests with real users: if a decision were about you, would the explanation pass the straight face test?
  • Setting bias checks on a fixed cadence: testing outcomes across groups you care about, recording the gaps, and documenting mitigations
  • Keeping a model change log in business language: what changed, why it changed, and what happened next

Regulation is moving in the same direction too. The EU and UK are both signalling that higher risk uses of AI should come with clear documentation, traceability, and routes to redress. International standards, such as ISO and NIST frameworks, turn that intent into practical control sets. None of this demands perfection. It asks for evidence that you know how your system behaves and that you can correct it when things go wrong.

The point is simple. Explainability is not a compliance chore to be done at the end. It is the way to unlock value at the start. When reasons are visible, the right people can challenge them. When challenges are welcomed, models improve. When models improve, customers benefit, colleagues trust the tools, and leaders can stand up in front of a room and say, with a straight face, why this system deserves its place in the business.

If you are adopting explainable AI, share what has worked, what has not, and the one change that made the biggest difference. Just drop us a line and we will welcome a discussion - info@meritosconsulting.com

Join 100+ subscribers

Get the latest news, blogs, and resources in your inbox monthly.

Subscribe Now
Your data is in the safe hands. Read our Privacy policy.

Thanks for subscribing!

You should receive a confirmation email shortly and we'll now send you new Perspectives as soon as they are published.

Oops! Something went wrong while submitting the form.