Generative AI (gen AI) promises to usher in an era of transformation for quality, accessibility, efficiency, and compliance in the financial services industry. As with any new technology, it also introduces new complexities and risks. Striking a balance between harnessing its potential and mitigating its risks will be crucial for the adoption of gen AI among financial institutions.
Historically, regulators and the financial services industry have developed various model risk management (MRM) frameworks to address the potential risks that arise from the use of models in decision-making. These principles-based frameworks typically encompass:
Our previous paper, written in partnership with the Alliance for Innovative Regulation (AIR) sought to assess the relevance of MRM for AI and machine learning (ML) models. Our latest joint paper expands on that foundation by exploring how MRM frameworks and established governance practices can be applied to manage risks in gen AI contexts.
Specifically, the paper proposes that regulators acknowledge best practices, provide enhanced regulatory clarity, and establish expectations in the following four areas: 1) model governance; 2) model development, implementation and use; 3) model validation and oversight; and 4) shared responsibility in third-party risk management.
Gen AI has the potential to contribute significantly to the economy, with estimates suggesting an addition of up to $340 billion annually to the banking sector alone. Financial institutions are already taking advantage of gen AI-based solutions to enhance efficiency, increase productivity among employees, improve customer engagement, and mitigate fraud and security risks.
Gen AI distinguishes itself from traditional AI by moving beyond analysis and prediction to creating new content. These models utilize probabilistic assessments, meaning they don’t produce a single definitive output but rather a range of possibilities based on the patterns they’ve learned. This capability unlocks new potential for human-computer interaction, allowing for more dynamic and creative applications.
While gen AI applications offer significant potential benefits, the technology also has unique characteristics and risks that should be assessed and mitigated. Importantly, existing MRM frameworks, designed to ensure the reliability and transparency of financial models, are flexible enough to accommodate gen AI deployment within financial institutions.
To help mitigate uncertainty as to how model risk can be managed to account for these unique aspects of gen AI, regulators could anchor to industry best practices and standards that they consider strong – perhaps presumptive – evidence that the requirements of MRM frameworks have been met.
Our new paper posits that clear governance frameworks that define roles, responsibilities, and accountability will be essential for effective oversight of gen AI. We highlight three key topics where additional regulatory clarity can benefit all stakeholders:
In this process, collaboration between industry participants, regulators, and governmental bodies will be key. While the path forward involves navigating complex regulatory and ethical landscapes, the collective commitment to responsible innovation and adherence to robust model risk management practices will be pivotal in realizing the full potential of gen AI in financial services and beyond.
Our new AI system accurately identifies errors inside quantum computers, helping to make this new…
Estimating the density of a distribution from samples is a fundamental problem in statistics. In…
Swiss Re & PalantirScaling Data Operations with FoundryEditor’s note: This guest post is authored by our customer,…
As generative AI models advance in creating multimedia content, the difference between good and great…
Large language models (LLMs) give developers immense power and scalability, but managing resource consumption is…
We dive into the most significant takeaways from Microsoft Ignite, and Microsoft's emerging leadership in…