Banks Are Trapped: Massive Compliance Spending Yet 98% of Criminal Money Goes Undetected
The global financial system is caught in a structural trap.

image from FLUX 2.0 Pro
The global financial system is caught in a structural trap. Banks and financial institutions spend 10 to 15 percent of their total headcount on Know Your Customer and Anti-Money Laundering compliance, yet they detect only about 2 percent of global illicit financial flows. That delta — massive investment against minimal return — is what Sam Boboev, a fintech analyst and the author of a deep-dive report published this week on Finextra, calls the compliance trap. His argument: the only exit is agentic AI, systems that plan, execute, and adapt sequences of actions autonomously — the difference between a chatbot that writes a summary and a digital worker that investigates a case end-to-end.
The scale of what institutions are losing has become harder to ignore. According to the Nasdaq Verafin 2026 Global Financial Crime Report, published March 11, 2026, illicit financial activity surged to an estimated $4.4 trillion in 2025, up $1.3 trillion in just two years — a compound annual growth rate of 19.2 percent that outpaces the world economy. Fraud scams and bank fraud together caused $579.4 billion in losses globally, with fraud scams growing at 19.3 percent — nearly double the rate of traditional bank fraud, according to the same report, which was produced in collaboration with Celent and Oliver Wyman and based on a survey of more than 500 financial crime professionals.
Criminal networks are not waiting for the financial industry to figure this out. According to the Interpol 2026 Global Financial Fraud Threat Assessment, AI-enhanced fraud is 4.5 times more profitable than traditional methods. Agentic AI systems can autonomously plan and execute complete fraud campaigns — from reconnaissance to ransom demands — without human intervention at any step. The implication is not subtle: the industry is being outpaced by adversaries who have already adopted the very technology that compliance teams are still piloting.
The architecture of the response matters. Boboev describes a squad of specialized agents, each mirroring a role along the compliance value chain. Retrieval-Augmented Generation agents extract data from profit-and-loss statements and beneficial ownership documents. Data pipeline agents orchestrate ETL processes and resolve entities across fragmented datasets. Research agents monitor market trends and counterparty patterns. Validation agents review agent outputs before anything reaches a human reviewer. The evaluation layer uses what Boboev calls LLM-as-judge: a secondary, more capable model assesses the primary agent output, flagging inconsistencies before human review. Citation validation ensures the system verifies that agent claims are grounded in retrieved data rather than model inference. The default remains human-in-the-loop for final dispositions — analysts approve, modify, or override the agent package, preserving accountability.
Early deployments are producing measurable results. Underdog Fantasy, an online fantasy sports and sports betting platform founded in 2020, cut its alert backlog by 72 percent after deploying Unit21 agentic platform, bringing its open queue from more than a thousand alerts to under 300 at any given time. Rule deployment time fell from two weeks to under five minutes, and built-in adverse media checks replaced manual processes that once consumed hours per case. The company AML Ops Lead, Dariusz Kaczmarek, described the shift: before Unit21, analysts were buried under false positives; today they focus on alerts that represent real risk. Nexo, a crypto lending platform, reduced false positives by 57 percent and is targeting 80 percent with the same platform. These results are not theoretical — they represent operational baselines that other institutions can evaluate.
The harder story is why most early agentic deployments fail. Boboev identifies three distinct failure modes. The hallucinating investigator occurs when teams provide too much context and open-ended prompts; in adversarial environments, the model fills data gaps with plausible but incorrect narratives. The over-suspicious agent results from pattern-driven training without contextual grounding — flagging high-value payments between related internal accounts as layering. The black box agent produces accurate outputs that cannot be defended to regulators because the reasoning chain is opaque. The common thread: poor guardrails, not weak models.
The deepest technical challenge is what Boboev calls context engineering. Large language models are based on transformer architecture, where every token attends to every other token, producing n-squared computational relationships. As context windows grow, attention becomes scarce — the model cannot meaningfully weight all available information. Effective context engineering is the discipline of curating high-signal tokens to maximize the likelihood of a correct outcome. Unit21, which has accumulated seven years of human review data, uses its dataset to determine the optimal context required for given tasks, evaluating agent outputs against historical investigations completed by high-performing analysts.
The regulatory landscape is shifting from skepticism to expectation. On March 6, 2026, the White House released its National Cyber Strategy and an Executive Order focused on combating cybercrime, fraud, and predatory schemes, explicitly encouraging AI-powered cyber defense to scale network protection. FINRA 2026 oversight report includes a dedicated AI section, emphasizing that firms must explain how AI is used and how outputs are tested. Agentic systems raise the stakes because they take actions, not just generate content — transparency into assumptions and accountability for outcomes are now non-negotiable requirements, not aspirational ones. The Financial Action Task Force has approved new strategic publications on cyber-enabled fraud and virtual asset risks, while Europe AMLA, the new anti-money laundering watchdog, is pushing banks from periodic customer reviews toward continuous understanding across the full customer lifecycle.
Criminal agents are already operating at speed. Boboev documents seven specific AI-driven fraud vectors now active in 2026. Agents embedded in compromised inboxes analyze historical threads and insert themselves into high-trust conversations, mirroring victims typing rhythms and working hours to bypass reputation-based security. Synthetic identity programs build credit histories programmatically over 18 months, then activate coordinated draining events across multiple institutions simultaneously. AI-enhanced fraud is growing substantially — Nasdaq Verafin documents the acceleration across multiple attack vectors, as real-time APIs power voice cloning and video manipulation that bypass biometric liveness checks once considered reliable. Laundering agents fragment stolen funds into tens of thousands of micro-transactions under $10, making the economic cost of manual tracing higher than the value recovered.
The structural argument is coherent and the primary source evidence is solid. What remains underspecified is the 20-fold productivity claim — a figure Boboev presents as an extrapolation from architectural potential, not a number measured in production. It is directionally consistent with the case study data from Underdog Fantasy and Nexo, but readers evaluating vendor claims in this space should know the distinction.
The path forward, as Boboev describes it, is pilot perimeters to prove impact before preparing for full-scale rollout. Agentic AI is not a product feature — it is an operating model change, and the institutions that get it right will have fundamentally reshaped how their compliance functions operate. The cost of staying in the trap is not abstract: $4.4 trillion in annual illicit activity is the measure of what the current system fails to catch.

