The banking world and “fast-moving innovation” have rarely been happy bedfellows. For decades, the financial sector has operated on a simple principle: stability over speed. But now, Artificial Intelligence is kicking down the door, and the old rulebook is looking decidedly dog-eared. Banks are champing at the bit to deploy AI for everything from spotting fraudsters to making smarter lending decisions, but they’re caught in a web of rules designed for a pre-AI era.
This is the very definition of bank AI adoption regulatory friction. It’s not that regulators are anti-AI; it’s that the frameworks are clunky, slow, and ill-suited for the rapid, iterative nature of modern technology. The result? A stalemate where innovation is stifled, and as we’ll see, our financial systems become less secure. Into this fray steps the U.S. Treasury Department, not with a shiny new law, but with something far more practical: a conversation. And it’s about time.
The Treasury’s Big Idea: Let’s Just Talk
When a government department announces a series of roundtables, it’s easy to be cynical. Is this just another talking shop destined to produce a dusty report? Perhaps. But the Treasury’s ‘AI Innovation Series’ feels different. Spearheaded by a collaboration between the Financial Stability Oversight Council (FSOC) and the new Artificial Intelligence Transformation Office (AITO), this initiative seems to grasp the core of the problem.
As reported by PYMNTS, the goal is to bring everyone to the table: the banks, the tech companies building the AI, and the regulators trying to keep up. This isn’t about the Treasury dictating terms from an ivory tower. It’s an admission that they don’t have all the answers. Paras Malik, the Treasury’s chief AI officer, put it plainly, stating the series aims to “ensure governance frameworks evolve alongside deployment and remain fit for purpose.” In other words, they’re trying to build the plane while flying it, and they need the pilots and engineers in the room.
It’s Not Deregulation, It’s Smart Regulation
Let’s get one thing straight. The push to reduce friction isn’t a call for a regulatory bonfire. The goal is optimisation. Think of it like this: inspecting a new car model for safety is essential. But if the rules demand that inspectors manually check every single screw after the car is fully assembled, you’re not making the car safer; you’re just ensuring it never leaves the factory. The current state of financial AI regulation often feels like that—a process so cumbersome it discourages the very innovation it’s meant to oversee.
This friction has real-world consequences. As Deputy Assistant Secretary Christina Skinner noted, “When institutions cannot deploy tools that improve fraud detection, credit allocation and operational resilience, the system becomes less efficient and less secure.” It’s a paradox: by being overly cautious with new tools, regulators inadvertently cling to older, less effective systems, leaving the door open for the very risks they fear. The real challenge is creating a regulatory MOT test that is rigorous but also designed for the F1-speed of banking technology innovation.
Building Bridges Before the Flood
The success of this entire endeavour hinges on genuine collaboration. For too long, these groups have operated in silos:
– Financial Institutions: Understand risk and compliance but are often slow to adopt new tech.
– Technology Companies: Move at lightning speed but can be naive about the systemic risks in finance.
– Regulators: Tasked with protecting the system but are often playing catch-up with the technology.
The AI Innovation Series represents a formal effort to bridge these divides. By getting these different worlds to speak the same language, the Treasury is tackling the problem at its root. This isn’t just about policy; it’s about culture and communication.
Creating a Common Language for AI
You can’t regulate what you can’t define. One of the most significant, yet overlooked, barriers to bank AI adoption regulatory friction is a simple lack of a shared vocabulary. Is an algorithm’s “bias” the same thing as a loan officer’s “discretion”? What constitutes “explainability” when an AI model has billions of parameters?
To solve this, the Treasury has released two foundational resources:
1. An AI Lexicon: This is more important than it sounds. It’s a Rosetta Stone for finance and tech, creating standardised definitions for key AI concepts. When a bank, a developer, and a regulator all mean the same thing when they say “model drift,” you’ve suddenly eliminated a huge source of misunderstanding and friction.
2. A Financial Services AI Risk Management Framework: This provides a practical toolkit for institutions. It’s designed to help them assess organisational readiness and manage AI risks internally, giving them a clear structure for building and deploying AI in a way that aligns with regulatory expectations for AI compliance.
These aren’t top-down mandates. They are foundational tools designed to empower the industry to innovate responsibly. It’s the difference between the government handing you a fish and teaching you how to build a better fishing rod.
Where the Rubber Meets the Road: Fraud and Lending
So, what does this all mean for you and me? Two of the most immediate applications for AI in banking are in fraud detection and credit allocation. AI algorithms can analyse patterns across millions of transactions in real-time, spotting sophisticated scams that would be invisible to human analysts. This doesn’t just save banks money; it protects consumers and strengthens the integrity of the entire payments system.
In credit, AI promises to make lending fairer and more accessible. By looking beyond simple credit scores to a wider array of data, AI models could potentially identify creditworthy individuals who are overlooked by traditional models. Of course, this is where the risks of bias are most acute, which is precisely why a clear and collaborative regulatory framework is so vital. Getting this right means unlocking capital for more people; getting it wrong could entrench existing biases on a massive scale.
The Bigger Picture: AI and Economic Security
This initiative isn’t just about making banking more efficient. As Treasury Secretary Scott Bessent articulated, it’s about preserving “U.S. financial sector’s leadership in AI adoption while preserving national security and long-term economic resilience.” This is the grand strategy at play. In a world where global economic power is increasingly tied to technological leadership, falling behind in the AI race isn’t an option.
If the U.S. financial system becomes a technological backwater, bogged down by antiquated rules while others race ahead, the consequences could be severe. Capital and talent will flow to more innovative ecosystems, eroding the U.S. dollar’s dominance and weakening the country’s strategic position. The Treasury Department AI push, therefore, is as much a matter of foreign policy as it is domestic regulation.
The future isn’t about whether banks will use AI. They will. The only question is whether they will do so within a framework that fosters responsible innovation or if they will be forced to choose between breaking the rules and becoming irrelevant. The Treasury’s approach suggests a welcome shift towards the former. It’s a pragmatic, collaborative, and long-overdue step. But the hard work is just beginning.
What do you think? Is this initiative the key to unlocking AI in finance, or will old habits die hard? Let me know your thoughts in the comments below.


