Banking Technology

Foundation Models Don’t Decline Transactions

Anthropic just shipped 10 financial services AI agents. Here’s why that validates the case for real-time fraud decisioning infrastructure — not replaces it.

RTD

RTD Team

Run-True Decision

Foundation Models Don’t Decline Transactions

The question is legitimate. When Anthropic announced ten purpose-built financial services AI agents on May 5, alongside Claude Opus 4.7 and a $1.5 billion joint venture with Blackstone, Goldman Sachs, and Hellman & Friedman, every fraud and risk leader in banking had a reasonable thought: does this change the calculus on specialised fraud decisioning platforms?

It changes something. But probably not what you are worried about.

What the Announcement Actually Covers

Anthropic’s ten new financial services agents target workflow-heavy, document-centric problems that analysts already spend their days on: KYC customer due diligence, AML case investigation, SAR narrative drafting, financial statement audit, month-end close reconciliation, and pitchbook preparation. The company’s partnership with FIS goes further — co-designing a Financial Crimes AI Agent that compresses AML alert investigations from days to minutes, with BMO and Amalgamated Bank already in development ahead of a general availability target in the second half of 2026.

These are real problems worth solving. AML investigation is labour-intensive, analyst-dependent, and notoriously slow. A complex transaction chain involving a suspected mule network can take a senior investigator several days to document, synthesise across data sources, and write up for compliance review. SAR narratives take hours to draft. KYC case assembly requires pulling entity data from multiple systems and packaging it for review by someone who did not build the original file. These workflows tolerate latency measured in seconds or minutes. They benefit from probabilistic, contextual reasoning. They involve human judgment and review at every material step. This is precisely where large language models perform well — and Anthropic has correctly identified it.

The announcement is credible, the partnerships are real, and the direction is right. Banks that dismiss it are behind. The productive question is not whether this matters, but what it actually displaces — and what it does not.

What Real-Time Fraud Decisioning Actually Requires

Real-time payment fraud operates under a different set of physics, and the physics are not negotiable.

When a customer initiates a transfer over PromptPay in Thailand, completes a QRIS scan at a merchant terminal in Indonesia, or sends an InstaPay transfer in the Philippines, the fraud decisioning engine has a window measured in milliseconds to evaluate the request, score the risk, apply configured detection logic, and return a decision — before the payment rail times out, and before the customer experience degrades. That window is not a design preference. It is an infrastructure constraint imposed by the payment rails themselves and the bank’s own channel commitments.

The capabilities required to operate within that window are specific. Deterministic execution: the same transaction inputs must produce the same decision, reliably, every time — because the governance model depends on it, and the regulator expects it. High-throughput scoring: a mid-sized bank processing real-time payments can see thousands of concurrent events per second during peak periods. On-premise deployment: transaction data in most Southeast Asian banking jurisdictions cannot traverse a third-party cloud inference endpoint — the Monetary Authority of Singapore’s Technology Risk Management Guidelines and similar frameworks in Indonesia, Thailand, and the Philippines are designed to support data residency requirements that on-prem deployment addresses directly. Full decision auditability: when an internal auditor asks why a specific transfer was declined on a specific date, the answer must be reproducible. Version-controlled governance: every rule and model change must be attributable, timestamped, and reversible.

Large language models are not structurally suited for inline transaction scoring. Inference latency is variable — a property of the architecture, not a tuning problem. The inference cost of running a large language model call for every payment transaction at meaningful scale is prohibitive at current market rates. And the probabilistic, non-deterministic nature of generative models is precisely the wrong property for a system where determinism is a hard requirement: the same input must produce the same output, in every jurisdiction, every time, for governance to function.

Latency variability and non-determinism are architectural properties, not performance shortfalls awaiting a faster GPU. Inference cost at scale is a real constraint today and will compress over time — but the structural incompatibility with deterministic governance requirements does not change with cost. Banks evaluating their fraud stack should hold this distinction clearly, because vendors that blur it are selling something the architecture cannot deliver.

Two Different Layers of the Same Stack

The most clarifying frame is architectural. A modern bank’s fraud operations stack has distinct layers, and AI investigation agents and fraud decisioning engines operate at different ones — which means the relationship is complementary, not competitive.

At the execution layer — where transactions are scored and decisions are made in real time — the requirements are determinism, low latency, high availability, full auditability, and governance. A fraud decisioning engine lives here. It processes the event stream, applies configured detection logic, orchestrates rules and model outputs, returns a decision, and logs an immutable record of what it saw and why.

At the investigation and operations layer — where analysts work queued alerts, build AML cases, draft compliance narratives, and conduct post-incident reviews — the requirements shift. Latency is measured in minutes, not milliseconds. The workflow is human-in-the-loop by design. The output is a document, a case file, a recommendation, not a real-time binary. This is where Anthropic’s financial services agents operate, and where they add genuine value.

These layers are not substitutes — they depend on each other. A bank deploying an AML investigation agent still needs real-time scoring underneath it. The investigation agent needs scored events, alert context, and a structured decision record to investigate. The fraud decisioning engine generates those inputs: the alert, the scoring rationale, the audit record, the version of the rule that fired. The AI investigation agent consumes and enriches that output, supports the analyst in building the case, and surfaces a richer picture for compliance review.

The future architecture is not a choice between AI agents and fraud infrastructure. It is AI agents operating above fraud infrastructure, making the compliance workflow faster and the analyst’s job less mechanical — while the execution layer continues to do the thing only it can do.

The Infrastructure Layer That AI-Native Banking Needs

Anthropic’s announcement is a meaningful signal about where enterprise banking is heading: toward AI-native operational workflows, natural-language interfaces, and embedded intelligence at every layer. Banks that deploy AI investigation agents will soon expect their fraud infrastructure to interoperate with that layer — to surface structured, explainable decisions that AI agents can consume, to support natural-language queries against historical decision data, to make rule configuration accessible to fraud operations analysts who are not engineers.

This is the direction that real-time fraud decisioning infrastructure needs to evolve toward. Not to become a language model — the execution requirements rule that out — but to become the reliable, auditable, interoperable foundation that AI agents and copilots can build on. LLM-assisted rule authoring: describe a detection pattern in plain language and get a draft rule back, ready for review and deployment. Natural-language audit replay: ask questions of a historical decision and receive a plain-language explanation of the inputs, the logic, and the outcome. Investigation copilot integration: surface structured context from the decisioning engine directly into the analyst’s AI-powered workflow, so the investigation starts from a richer starting point than a bare alert ID.

The engine stays the execution anchor. AI makes it faster to configure, easier to audit, and more accessible to the operational team running it day to day.

Anthropic’s move into financial services is not evidence that the fraud execution layer is becoming redundant. It is evidence that banks are investing in AI operations at scale — and that the infrastructure layer those operations depend on matters more, not less, as that investment grows. The question fraud and risk leaders should be asking is not “does this replace our decisioning platform?” It is “how does our decisioning platform connect to the AI agents we are going to deploy?”

Run-True Decision is building a fraud decision engine purpose-built for Southeast Asian banks — designed to serve as the real-time risk infrastructure layer for AI-native banking operations. Talk to us if your team is thinking through how AI agents fit into your fraud stack.

Explore the Platform

See how Run-True Decision handles real-time fraud scoring, on-premise deployment, and regional compliance for Southeast Asian banks.

View Platform Overview

Related Articles