top of page

Why Your AI Agents Can't Pass an Audit (And How to Fix It)

You've deployed an AI agent to handle loan pre-screening. It processes hundreds of applications a day, faster and cheaper than any manual team. Leadership is happy. Then your compliance officer gets a call from the regulator.


They want to know why a specific application was declined.


You pull the logs. You find the input. You find the output. But between those two points - the reasoning, the logic, the actual decision path: there's nothing. Just a number from a model that no longer runs the same way it did last Tuesday.


That's not an audit trail. That's a black box with timestamps.


This is the compliance gap growing quietly inside most enterprise AI deployments. And as the EU AI Act moves into full enforcement from August 2026, with penalties reaching €35 million or 7% of global annual turnover, it's about to become a very expensive problem.




The audit illusion


Most organisations deploying AI agents today believe they're covered because they're logging. They capture inputs, outputs, timestamps, maybe token counts. They have dashboards. They have monitoring.


But logging what an AI said is not the same as being able to explain why it said it.

Regulators (under the EU AI Act, under Basel IV, under sector-specific frameworks) are not asking for output records. They're asking for decision logic.


They want to know:

  • What rules governed this decision?

  • Could you reproduce this outcome with the same inputs tomorrow?

  • Who approved the logic that produced this result?

  • Could a non-technical stakeholder verify it?


If your AI agent is making decisions through a large language model, reasoning its way to an answer at runtime, you cannot reliably answer any of those questions.



The root cause: Probabilistic systems can't make auditable promises


LLMs are probabilistic by design. That's not a flaw; it's what makes them powerful. The same capability that lets them reason across complex contexts, generate nuanced responses, and adapt to novel inputs also means that identical inputs can produce different outputs across runs, prompts, or model versions.


For creative tasks, that's a feature. For regulated decisions, it's a liability.


Consider what happens when an LLM is responsible for executing business logic directly: a credit rule, a risk classification, a compliance check. The model may apply the right logic most of the time. But "most of the time" is not a standard that holds up in front of a regulator. And you cannot show an auditor the reasoning path of a system that generates its logic dynamically each time it runs.


This is what compliance teams are starting to discover: the problem isn't that their AI makes bad decisions. The problem is that they can't prove their AI makes consistent, rule-governed ones.




What a real audit trail requires


Before fixing the problem, it's worth being precise about what auditability actually demands in practice. Three things are non-negotiable:


Traceability. Every decision must have a reproducible chain of logic — not a statistical approximation of how an answer was reached. Article 12 of the EU AI Act requires high-risk AI systems to automatically log events throughout their lifecycle in a way that enables post-market monitoring and risk identification. If the same inputs are submitted again, the system must produce the same output, and you must be able to show the path from one to the other.


Human oversight. Under Article 14 of the EU AI Act, organisations deploying high-risk AI systems must ensure that a human can understand, monitor, and intervene in AI-driven decisions. That's structurally impossible if the decision logic lives inside a black-box model that only engineers can partially interpret.


Transparency. Article 13 mandates that high-risk AI systems be designed so that deployers can interpret outputs and use them appropriately — with clear instructions on capabilities, limitations, and how to collect, store, and interpret logs. "The model decided" is not a compliant answer.


Logging gives you none of these. Deterministic decision logic gives you all three.




An example: SME loan eligibility at a bank


Consider a mid-sized bank running an AI-assisted lending workflow for small business loans. Their AI agent handles document processing, customer communication, and data extraction. It's efficient, it reduces manual workload significantly, and the business unit loves it.


The problem surfaces when the compliance team runs an internal audit ahead of the EU AI Act deadline.


They select ten declined applications from the previous quarter and ask a straightforward question: given the same applicant data, would the system make the same decision today?

In three out of ten cases, the answer is uncertain. The model has been updated twice. The prompts have been refined. There's no fixed logic that can be re-run against historical inputs to reproduce the original output. The decision effectively no longer exists in any verifiable form.


When the compliance officer asks the engineering team to explain the reasoning behind one specific decline, a borderline case where the applicant is now disputing the outcome, the honest answer is that they can't. They can show what data went in and what came out. The middle is opaque.


This is not a hypothetical. 89% of financial institutions are actively pursuing GenAI adoption. Most are doing so without the governance infrastructure to defend the decisions those systems make.


What the fix can look like:


The bank separates its AI stack into two distinct layers. The AI agent continues to handle flexible tasks: document extraction, customer interaction, gathering the inputs needed to make a decision. But when it reaches the actual eligibility determination, it no longer reasons its way to an answer. Instead, it calls a predefined logic blueprint.


That blueprint contains the bank's credit policy encoded as a deterministic, visual decision model: built by the risk team, reviewed by compliance, approved by legal, and locked into version control. The inputs go in. The rules execute. The same output comes out every time. Every execution is logged against the specific version of the logic that ran.


Now when a regulator asks why an application was declined, the compliance team can open the blueprint for that loan, run through the decision path step by step, and show exactly which condition was not met; in a diagram that anyone in the room can follow without a computer science degree.


That's not just compliance. That's confidence.



The architecture fix


The core architectural principle here is separation of concerns. AI models are excellent at reasoning, interpreting context, and handling the unpredictable. They are structurally unsuited to be the source of truth for business rules.


The solution is to keep LLMs where they belong — flexible reasoning and orchestration — and route critical decisions through deterministic logic that is defined, validated, and stored independently.


This means:


  • Decision logic is designed before it runs, not generated at runtime. Domain experts and compliance teams build the rules in advance, in a form they can read and approve.

  • Blueprints are executed, not prompted. When an AI agent needs to make a consequential decision, it calls a logic tool that executes the pre-approved rules. No LLM in the decision path.

  • Every execution is traceable. The same input will always produce the same output, with a full record of which version of the logic ran and when.

  • Non-technical stakeholders can verify it. Visual blueprints mean compliance officers, legal teams, and auditors can inspect the logic themselves — without needing to read code.


This is what "auditable by design" actually means. Not a compliance feature bolted on after the fact. An architecture built from the ground up so that explainability is the natural output.




The business case beyond the audit


It's worth noting that the value here isn't purely defensive. According to McKinsey's 2025 State of AI research, CEO oversight of AI governance is one of the factors most correlated with higher bottom-line impact from generative AI. The organisations getting more value from AI are not treating governance as an afterthought — they're building it into how their systems operate.


When logic is visual and explicitly versioned, business teams can update rules directly without going through a development backlog. Policy changes that previously took weeks of back-and-forth between compliance and engineering can be made, reviewed, and deployed in hours. The compliance infrastructure and the operational efficiency gain are the same architecture.



The clock is running


The EU AI Act is fully applicable from August 2, 2026. For high-risk AI systems, including those used in credit scoring, insurance underwriting, and employment decisions, the requirements for traceability, human oversight, and documentation are not optional.


Most organisations still have time to act. But the window is narrowing, and technical debt accumulates quickly when AI agents ship faster than governance infrastructure can keep up.

The question for compliance and technology leaders isn't whether to build an auditable AI architecture. It's whether you build it before or after your first uncomfortable conversation with a regulator.



Explore Leapter with an expert


Leapter builds the deterministic logic layer for enterprise AI agents; visual blueprints that give compliance teams the audit trail they need and domain experts the control they deserve.



 
 
bottom of page