The use of technology does not excuse unlawful discrimination, according to federal regulators. But as artificial intelligence (AI) models make their way into underwriting, they introduce the potential for discrimination that fair lending laws and banks’ compliance practices aren’t designed to handle. AI’s capabilities and the murkiness of models may get in the way of lending decisions that are sound, legal, and explainable. Regulators are watching: In 2023, four federal agencies published a statement about enforcing fair lending laws and regulations in the age of AI. Their bottom line: Complexity and opacity aren’t a defense against violations of fair lending regulations.
A series of laws first passed in the 1960s addressed a problem of an age long past: One in which limited information to inform loan decisions allowed too much discretion to loan officers. Today’s problem is the opposite: As the Bank Policy Institute argues, there is too much data, and regulations based on fair lending laws were not designed with today’s wide availability of demographic and behavioral data and raw computing power in mind.
Even with objective analysis based on a lot of data, bias exists as “disparate impact.” That occurs when an underwriting model has a “disproportionately adverse impact on applicants from a protected class, unless the practice meets a legitimate business need that cannot reasonably be achieved by means that are less disparate in impact,” according to the Bank Policy Institute.
As the banking industry has computerized, it’s made do with computational underwriting models that account for that compliance challenge. Models have reportedly been simple enough to be explainable, generate the legally required rationale for credit decisions, and allow banks to test systems for compliance. But AI poses challenges to that status quo because AI-based credit underwriting systems do not always transparently or accurately provide the specific reasons for adverse credit decisions, putting them out of compliance.
As lenders assess challenges with AI-based underwriting, they have two big issues to keep in mind, according to the Consumer Financial Protection Bureau (CFPB):
Black box algorithms: Fair lending laws require banks to explain “the specific reasons for denying an application for credit or taking other adverse actions, even if the creditor is relying on credit models using complex algorithms.” If nobody knows exactly how the model works, including its developers, it may not be possible to comprehend the algorithm’s decision-making process or determine its biases.
Digital redlining: Statistical models can infer characteristics protected by fair lending laws, and intentionally or not, those characteristics can be reflected in underwriting decisions. As the Brookings Institution argues, there are “too many sources of data that can hide as proxies for illegal discrimination.” The Equal Opportunity Credit Act specifically prohibits discrimination based on “race, color, religion, national origin, sex, marital status, age” and several other factors.
Despite real issues, the ability to evaluate more data for credit decisioning using AI systems with the appropriate guardrails is an opportunity for lenders. Traditional credit scoring and reporting has its own biases, including potential borrowers’ need for a credit history to acquire more credit. Holistic analysis using AI models that includes “alternative” credit scoring data made available through open banking, for example — like bill and rent payments and spending patterns — may make credit scoring fairer. But solving the problem of bias in AI systems will need to happen first. That, ultimately, will require solving for bias in data the models are trained on — which itself is complicated to balance.
We use cookies on our website to give you the most relevant experience by remembering your preferences and repeat visits. By clicking “Accept”, you consent to the use of ALL the cookies.
This website uses cookies to improve your experience while you navigate through the website. Out of these, the cookies that are categorized as necessary are stored on your browser as they are essential for the working of basic functionalities of the website. We also use third-party cookies that help us analyze and understand how you use this website. These cookies will be stored in your browser only with your consent. You also have the option to opt-out of these cookies. But opting out of some of these cookies may affect your browsing experience.
Necessary cookies are absolutely essential for the website to function properly. This category only includes cookies that ensures basic functionalities and security features of the website. These cookies do not store any personal information.
Any cookies that may not be particularly necessary for the website to function and is used specifically to collect user personal data via analytics, ads, other embedded contents are termed as non-necessary cookies. It is mandatory to procure user consent prior to running these cookies on your website.
A Problem With AI-Based Credit Models
A Problem With AI-Based Credit Models
February 6, 2024
By: Tyler Brown
Request for Proposal
By: Tyler Brown
The use of technology does not excuse unlawful discrimination, according to federal regulators. But as artificial intelligence (AI) models make their way into underwriting, they introduce the potential for discrimination that fair lending laws and banks’ compliance practices aren’t designed to handle. AI’s capabilities and the murkiness of models may get in the way of lending decisions that are sound, legal, and explainable. Regulators are watching: In 2023, four federal agencies published a statement about enforcing fair lending laws and regulations in the age of AI. Their bottom line: Complexity and opacity aren’t a defense against violations of fair lending regulations.
A series of laws first passed in the 1960s addressed a problem of an age long past: One in which limited information to inform loan decisions allowed too much discretion to loan officers. Today’s problem is the opposite: As the Bank Policy Institute argues, there is too much data, and regulations based on fair lending laws were not designed with today’s wide availability of demographic and behavioral data and raw computing power in mind.
Even with objective analysis based on a lot of data, bias exists as “disparate impact.” That occurs when an underwriting model has a “disproportionately adverse impact on applicants from a protected class, unless the practice meets a legitimate business need that cannot reasonably be achieved by means that are less disparate in impact,” according to the Bank Policy Institute.
As the banking industry has computerized, it’s made do with computational underwriting models that account for that compliance challenge. Models have reportedly been simple enough to be explainable, generate the legally required rationale for credit decisions, and allow banks to test systems for compliance. But AI poses challenges to that status quo because AI-based credit underwriting systems do not always transparently or accurately provide the specific reasons for adverse credit decisions, putting them out of compliance.
As lenders assess challenges with AI-based underwriting, they have two big issues to keep in mind, according to the Consumer Financial Protection Bureau (CFPB):
Despite real issues, the ability to evaluate more data for credit decisioning using AI systems with the appropriate guardrails is an opportunity for lenders. Traditional credit scoring and reporting has its own biases, including potential borrowers’ need for a credit history to acquire more credit. Holistic analysis using AI models that includes “alternative” credit scoring data made available through open banking, for example — like bill and rent payments and spending patterns — may make credit scoring fairer. But solving the problem of bias in AI systems will need to happen first. That, ultimately, will require solving for bias in data the models are trained on — which itself is complicated to balance.
You Might Like These, Too
Cloud Powers Broad AI Use Cases in Banking
A Quarter of Banks Are Ready for a New Core
A Year After SVB’s Fall, Cash Management Remains in Focus for Startups
The Clock Starts Soon for Open Banking Compliance
Leaders in Bank Consulting