Cabin logo
Drag
Play
Cabin
  • AboutAbout
  • ServicesServices
  • InsightsInsights
  • CareersCareers
MenuClose
  • ContactContact
  • About
  • Services
    • Strategy & Innovation
    • Product Design
    • Software Engineering
    • View all Services
  • Insights
  • Careers
  • Contact
Social
  • LinkedIn
Charlotte Office
421 Penman St Suite 310
Charlotte, North Carolina 28203
Get in touchGet in touch

Generative AI in Financial Services: What Actually Ships

March 18, 2026
|
16 min read
Cabin
Cabin

Last updated: March 2026

The generative AI use cases that get the most attention in financial services — the ones in every vendor deck and conference keynote — are rarely the ones that make it to production. Not because the technology isn’t capable. Because they were designed for the demo environment, not the regulated one.

The use cases that actually ship look different. Less flashy, more constrained, and built around a reality that most vendor content glosses over: in financial services, a wrong model output isn’t a bad user experience. It’s a compliance event, a fair lending violation, or a model risk finding that sends the whole project back to square one.

BCG’s 2024 AI adoption research found that 74% of companies still haven’t generated tangible value from AI at scale. In financial services, the picture is sharper: a 2025 BFSI sector survey found that while 52% of financial services organizations are using generative AI, only 8% are doing it strategically at the enterprise level. MIT’s research on generative AI pilots finds a 95% failure rate — and explicitly calls out that in regulated industries like financial services, in-house builds fail more often than anywhere else.

The gap isn’t ambition. It’s the distance between what gets approved in a proof-of-concept and what survives the model risk review.

This article is a practitioner’s breakdown of which generative AI use cases in financial services hold up — and why the ones that don’t fail in entirely predictable ways.

Why Most Generative AI Use Cases in Financial Services Stall

Most generative AI use cases in financial services stall not because the technology isn’t capable, but because they were designed for clean demo environments rather than regulated production ones. The failure modes are consistent: auditability gaps that surface in model risk reviews, data infrastructure that wasn’t built for LLM inputs, and human-in-the-loop requirements that weren’t designed in from the start.

The model risk management framework most US banks operate under — SR 11-7, now explicitly extended by OCC bulletins to cover AI and LLMs including third-party tools — requires that any model used in a material business decision be validated, documented, and monitored on an ongoing basis. Generative AI models don’t fit cleanly into traditional backtesting validation approaches: they’re probabilistic, context-sensitive, and difficult to validate the way a credit scorecard is validated. Teams that ship generative AI in financial services resolve that tension in the architecture before they start building. The ones that don’t find it in the compliance review — at which point it’s a rebuild, not a patch.

The data infrastructure problem is less discussed but equally common. Financial services organizations run on legacy systems that weren’t built to feed LLM pipelines. Getting clean, structured, permissioned data into a model at inference time is often 60-70% of the actual project work. That work doesn’t show up in the vendor demo and doesn’t appear in the initial project scope until someone starts the integration.

Human-in-the-loop requirements are the third failure mode. Financial services decisions with regulatory weight — credit decisions, AML flags, customer communications — require documented human review at points in the workflow that the system has to be designed around. An agentic system that acts autonomously on those decisions isn’t just a product risk. It’s a regulatory one. The systems that hold up are the ones where human checkpoint placement was a first-order design decision.

The Use Cases That Actually Make It: What They Have in Common

The generative AI use cases that reach production in financial services aren’t the most technically ambitious. After building these systems in financial services for years, here’s what the ones that ship have in common — and it’s not what most vendor roadmaps emphasize.

They produce structured, auditable outputs

The use cases that ship produce outputs a compliance team can audit and a model risk team can validate. Document summaries with cited sources. Extracted data fields with confidence scores. Flagged transactions with reasoning chains. Not free-form generation. Structured outputs where a good output can be defined, tested, and monitored over time. If you can’t define what a correct output looks like, you can’t validate the model that produces it under SR 11-7.

They keep humans in the loop at the right points

Production financial services AI doesn’t eliminate human judgment. It routes it. The use cases that hold up place human review at the decisions that carry regulatory weight and remove it from the decisions where automation is safe and volume makes human review unsustainable. That placement decision has to be made explicitly at architecture — not adjusted later when the compliance team asks about it.

They work with the data infrastructure that exists now

The use cases that ship were designed around the data that’s actually accessible, not the unified data lake on the three-year roadmap. That usually means starting with documents: contracts, filings, loan files. Documents are already in accessible formats. They don’t require live system integration to process, and they don’t need a data modernization project to unlock.

They were built for the model risk review from day one

The teams that get through model risk without a rebuild wrote the documentation outline before they finished the architecture. Validation approach, performance metrics, human oversight mechanisms, known limitations — all defined at design. An AI-native build process makes this faster: we’re running Claude Code in current financial services engagements to generate draft model documentation alongside the architecture decisions, so the compliance artifacts aren’t trailing the build by weeks.

Document Intelligence: The Use Case Most Teams Underestimate

If there’s one generative AI use case in financial services that consistently outperforms expectations in production, it’s document intelligence — and it’s consistently underestimated in the planning phase because it doesn’t sound as impressive as the use cases that lead vendor decks.

Contract analysis. Regulatory filing review. Loan file processing. Credit memo summarization. Earnings call analysis. These are the use cases financial services organizations are actually running in production, and for good reason: they fit the constraints that kill other use cases.

The outputs are structured and reviewable. A generative AI system that extracts key terms from loan agreements and flags exceptions against a policy checklist produces outputs a human reviewer can validate in minutes. That’s a model risk team’s preferred deployment pattern: AI as a first-pass reviewer, human as the decision-maker, audit trail at every step.

The data infrastructure requirement is manageable. Documents are accessible. They don’t require live system integration, real-time data feeds, or a unified data model. A financial services organization can run document intelligence on existing file repositories without rebuilding its data architecture first.

The ROI is measurable and fast. In our work, document intelligence workflows — loan file review, contract analysis, regulatory document processing — routinely deliver 5-10x throughput improvements over manual review. That’s consistent with what JPMorgan has described publicly with COiN, their contract intelligence system: hundreds of thousands of hours of manual commercial loan agreement review compressed into seconds. Throughput numbers vary by document type and complexity, but the direction holds across every production implementation we’ve run.

We’ve built document intelligence systems for financial services clients where the model risk submission cleared in the first review cycle — no rebuild, no resubmission. The reason: the architecture was designed around SR 11-7 requirements before a line was built. Validation approach defined upfront. Human review checkpoints at the extraction confidence threshold and at any flag triggering a policy exception. Output format structured so the audit trail is embedded in the output, not reconstructed from logs after the fact.

Document intelligence isn’t the use case that leads most conversations about generative AI in financial services. It’s the one that ships first, delivers measurable ROI fastest, and builds the internal model risk credibility that makes the more ambitious use cases approvable later. Start here.

Agentic Workflows in Financial Services: What Works and What Breaks

Agentic AI — systems that take sequences of actions autonomously to complete a task — is where most of the ambition in financial services AI is concentrated right now. It’s also where most of the production failures happen.

The use cases that hold up are the ones where the agent operates in a closed, well-defined environment with reversible actions and human review at the output:

Internal research and synthesis. An agent retrieves filings, synthesizes analyst reports, and produces a structured briefing for a human decision-maker. The agent handles the research phase autonomously. A human makes the decision. The output is reviewable and every action is logged.

Back-office workflow automation. Document routing, data extraction, exception flagging, status tracking. High volume, low regulatory stakes per transaction, reversible if something goes wrong. Agentic systems handle this well because the cost of an error is low and the throughput improvement is high.

Compliance monitoring and flagging. An agent that monitors communications or transactions for policy exceptions and routes flags to human reviewers. The agent identifies, the human decides. Audit trail at every step.

The use cases that break are the ones where the agent makes consequential decisions autonomously:

Use Case Why It Stalls
Autonomous credit decisioning Adverse action notice requirements, fair lending obligations — human review required by regulation
Customer-facing AI advisors Suitability requirements, fiduciary obligations — AI can inform but not advise without human oversight
Real-time AML decisioning SAR filing obligations require documented human judgment, not autonomous model decisions
Automated regulatory reporting Material errors in regulatory submissions carry severe penalties — human sign-off non-negotiable
Autonomous trading execution Market risk, operational risk, regulatory requirements — human oversight at execution is standard

The pattern is consistent: agentic systems work in financial services when the agent handles the volume work and a human handles the consequential decision. Systems that invert that pattern — agent makes the decision, human reviews the exception — run into regulatory requirements that weren’t designed with autonomous AI in mind.

Building the orchestration layer to support this correctly is where most teams underestimate the architecture work. For a deeper look at how to design the decision layer that makes agentic financial services systems hold up, see our orchestration layer guide and our approach to building AI systems people actually use.

Customer-Facing AI in Regulated Environments: The Constraints That Determine Success

Customer-facing generative AI in financial services is the use case that gets the most attention and has the longest path to production. The technology is capable. The regulatory environment is the constraint — and it’s more layered than most vendor roadmaps account for.

Three requirements shape whether a customer-facing use case is buildable:

Adverse action notice requirements. The CFPB has been explicit: lenders must provide specific, accurate reasons when taking adverse action on a credit application, and that obligation “remains even if those companies use complex algorithms and black-box credit models that make it difficult to identify those reasons.” A generative AI system that influences a credit decision has to produce documentable reasons — not boilerplate codes, not “the model said no.” Systems that can’t produce that explanation aren’t deployable in a credit context regardless of their accuracy.

Fair lending obligations. CFPB guidance makes clear that AI-driven decisions face the same fair lending analysis as any other decisioning model: disparate impact testing, monitoring for proxy discrimination, and documentation of training data and validation approach. Teams that don’t build fair lending analysis into the development process find it in the compliance review — at which point it requires a rebuild.

Suitability, fiduciary, and FINRA supervision requirements. For investment-related customer interactions, the SEC’s Investor Advisory Committee has recommended that any AI interacting directly with investors meet Regulation Best Interest and fiduciary duty standards. FINRA’s 2024 guidance makes it explicit that AI-generated communications — including chatbots — are subject to Rule 2210 on fair and balanced communications and Rule 3110 on supervision. If a bot talks to a customer, the firm owns the message and has to supervise and archive it like any other broker-dealer communication. AI can inform and present options. Recommendation outputs that carry regulatory weight need human review before they reach the customer.

None of these requirements make customer-facing financial services AI impossible. They make it more constrained than most vendor roadmaps account for. The systems that make it to production are the ones where the regulatory constraints were treated as design inputs from the start — not compliance boxes to check after the system was built.

For more on what an AI transition looks like when regulatory constraints are a first-order input, see our AI transition guide. For how to structure the engagement so the regulatory design stays inside your organization after it ends, see how we address the structural traps that create consultant dependency.

How to Evaluate a Generative AI Use Case Before You Build

The question most financial services teams ask too late: is this use case actually buildable in our regulatory environment, with our data infrastructure, on our timeline?

The evaluation framework we use before committing to a build:

Can the output be structured and audited? Free-form generative outputs are hard to validate under model risk frameworks. If the use case requires outputs that a model risk team can review, approve, and monitor — define the output structure before the architecture. If you can’t define what a good output looks like, you can’t validate the model that produces it.

Where does human judgment have to sit? Map every point in the workflow where a regulatory requirement, a risk policy, or a business rule requires documented human review. Those are your human checkpoints. If the use case requires human review at so many points that the automation benefit disappears, the use case isn’t ready — or the workflow needs to be redesigned before the AI is added.

What data does the use case actually need, and do you have it? Not the data you’ll have after the data modernization project. The data you have now, in the format it’s currently in. Document-based use cases are almost always more accessible than use cases that require live system integration. If the use case requires data that isn’t accessible yet, scope the data work before the AI work — or choose a use case that works with what you have.

Can you write the model risk documentation before you finish the build? If you can’t describe the validation approach, the performance metrics, the human oversight mechanisms, and the limitations of the model before the build is complete, the model risk submission will find those gaps. Write a draft documentation outline at architecture. If it’s hard to fill in, the architecture has gaps worth finding now.

What does the failure mode look like, and what’s the cost? For every use case, define the failure mode explicitly. Not “the model gets it wrong” — what does it get wrong, how often, and what’s the downstream impact? A document intelligence system that misses a contract clause has a different failure cost than a customer communication system that produces a non-compliant disclosure. Know the failure cost before you build.

For more on how this evaluation framework connects to a broader AI product development approach, see how we approach digital product development and how this fits into an AI transition your team can own.

Frequently Asked Questions

What generative AI use cases are financial services firms actually using in production?

The use cases with the highest production rates in financial services are document intelligence (contract analysis, regulatory filing review, loan file processing), internal research and synthesis, back-office workflow automation, and compliance monitoring with human review. Customer-facing AI and autonomous decisioning use cases are less common in production because of the regulatory requirements around explainability, adverse action notice, and fair lending.

How does model risk management affect generative AI deployment in financial services?

SR 11-7 and related model risk guidance require that any model used in a material business decision be validated, documented, and monitored. Generative AI models are harder to validate using traditional approaches because they’re probabilistic and context-sensitive. The teams that navigate this successfully define the validation approach, output structure, and human oversight mechanisms at architecture — before the model risk submission, not after the review finds gaps.

What’s the fastest generative AI use case to get to production in a financial services organization?

Document intelligence — contract analysis, loan file review, regulatory document processing — consistently reaches production fastest because it fits the regulatory environment most cleanly. Outputs are structured and reviewable, data is accessible without major infrastructure work, human review integrates naturally into the workflow, and the ROI is measurable from the first weeks of production. It’s also the use case most teams underestimate when they’re planning their generative AI roadmap.

Generative AI in financial services isn’t hard to pitch. It’s hard to ship. The use cases that make it aren’t the ones that look best in the demo — they’re the ones that were designed for the environment the demo never shows: the model risk review, the data infrastructure gap, the regulatory constraint that the vendor content doesn’t mention.

The organizations getting this right are the ones that treated those constraints as design inputs, not compliance hurdles to clear after the build.

If you’re evaluating generative AI use cases for a financial services engagement, we’ve shipped these — through model risk reviews, with compliance teams in the room, on data infrastructure that wasn’t built for AI. Worth a conversation.

About the Author

Cabin is an AI transformation consultancy that architects AI-native products, implements intelligent systems, and builds client team capability while doing it. Founded by the core team behind Skookum, which became Method under GlobalLogic and rolled up to Hitachi, Cabin’s partners have shipped 40+ enterprise products together over nearly 20 years, for clients including FICO, American Airlines, First Horizon, Mastercard, Trane Technologies, and SageSure.

Design system implementation is where Cabin operates every day, not as an advisor watching from the sidelines, but as the senior designers, engineers, and strategists doing the work. The team has built and rescued design systems across financial services, healthcare, and insurance — embedding with client teams, not above them, so the capability stays when the engagement ends.

Everything Cabin publishes on design systems, DesignOps, and team enablement comes from work currently in progress, not from research reports or conference decks. When we write about why design systems fail, it’s because we’ve inherited the aftermath. When we write about governance that works at scale, it’s because we’ve built the playbooks.

About the author
Cabin
Cabin

Related posts

  • AI Transition: Why Most Organizations Get It Wrong

    AI Transition: Why Most Organizations Get It Wrong

    March 18, 2026
       •   14 min read
    Cabin
    Cabin
  • Healthcare AI Consulting: What to Ask Before You Sign

    Healthcare AI Consulting: What to Ask Before You Sign

    March 18, 2026
       •   17 min read
    Cabin
    Cabin
  • Orchestration Layer: Where AI Products Break Down

    Orchestration Layer: Where AI Products Break Down

    March 18, 2026
       •   14 min read
    Cabin
    Cabin
  • AI
    AI Transition: What Actually Changes [Enterprise Guide]

    AI Transition: What Actually Changes [Enterprise Guide]

    February 27, 2026
       •   15 min read
    Cabin
    Cabin
  • AI
    AI Agents in Enterprise: What Actually Ships [2026]

    AI Agents in Enterprise: What Actually Ships [2026]

    February 27, 2026
       •   14 min read
    Cabin
    Cabin
  • AI
    LLM Integration Is Harder Than an API Call [What Teams Miss]

    LLM Integration Is Harder Than an API Call [What Teams Miss]

    February 27, 2026
       •   14 min read
    Cabin
    Cabin
  • Design
    Dashboard UX Best Practices That Drive Adoption

    Dashboard UX Best Practices That Drive Adoption

    February 19, 2026
       •   11 min read
    Brad Schmitt
    Brad Schmitt
  • Strategy
    Consultant Knowledge Transfer: What Actually Works

    Consultant Knowledge Transfer: What Actually Works

    February 18, 2026
       •   11 min read
    Brad Schmitt
    Brad Schmitt
  • Strategy
    Team Enablement Consulting: Skills, Not Dependency

    Team Enablement Consulting: Skills, Not Dependency

    February 18, 2026
       •   13 min read
    Brad Schmitt
    Brad Schmitt
  • Design
    How to Build a Design System Teams Actually Use

    How to Build a Design System Teams Actually Use

    February 15, 2026
       •   11 min read
    Brad Schmitt
    Brad Schmitt
  • Salesforce
    Salesforce Implementation Checklist: Complete Guide [2026]

    Salesforce Implementation Checklist: Complete Guide [2026]

    January 20, 2026
       •   9 min read
    Brad Schmitt
    Brad Schmitt
  • Strategy
    Team Capability Building That Actually Sticks

    Team Capability Building That Actually Sticks

    January 20, 2026
       •   10 min read
    Brad Schmitt
    Brad Schmitt
Logo
A digital experience consultancy powered by AI-driven innovation
→Get in touch→
  • Contact
    [email protected]
  • Social
    • LinkedIn
  • Charlotte office
    421 Penman St Suite 310
    Charlotte, North Carolina 28203
  • More
    Privacy Policy
© 2025 Cabin Consulting, LLC