AI for compliance: spotting where AI goes rogue

AI tools are rapidly becoming part of day-to-day compliance work in financial services, but a key issue is lack of confidence in their outputs. This raises questions around reliability, oversight, and defensibility. This workshop focuses on how to design AI-enabled compliance workflows that can be trusted in real operational settings.

As teams in financial services increasingly use AI tools like ChatGPT in their day-to-day work, a key challenge is how to ensure the outputs are reliable.

This workshop focuses on the design of human-in-the-loop compliance AI workflows that can be relied on in operational settings.

You’ll learn how to:

  • Design evaluation models that assess AI outputs for accuracy, completeness, and regulatory alignment
  • Define human review points and design interventions that meaningfully improve reliability
  • Use confidence scoring and guardrails to detect uncertainty, prevent drift, and constrain model behaviour
  • Design for repeatability so outputs remain consistent across users, prompts, and time

Daniel Chatfield, Chief Information Officer (EU) at Monzo

Daniel Chatfield, Chief Information Officer (EU) joins the session to discuss how Monzo used AI to support regulatory work during rapid growth, the initial efficiency gains achieved, and the key barriers encountered, including sourcing, version accuracy, and reliability.

He will focus on where human review and judgement were required, how outputs were checked and escalated, and the practical lessons learned for building repeatable and scalable AI-enabled workflows.

Who should attend

Compliance leaders, policy teams, MLROs, risk and assurance professionals, audit teams, and anyone responsible for designing or overseeing AI-enabled compliance processes.

Previous sessions and resources

Visit https://www.zango.ai/events for a recording of our previous session: "Prompt like an AI engineer", and accompanying resources.

Continue reading