{"id":1409461,"date":"2025-07-14T16:50:20","date_gmt":"2025-07-14T14:50:20","guid":{"rendered":"https:\/\/www.ie.edu\/insights\/?post_type=articles&#038;p=1409461"},"modified":"2025-07-14T16:50:20","modified_gmt":"2025-07-14T14:50:20","slug":"rethinking-ai-in-credit-decision-making","status":"publish","type":"articles","link":"https:\/\/www.ie.edu\/insights\/articles\/rethinking-ai-in-credit-decision-making\/","title":{"rendered":"Rethinking AI in Credit Decision-Making"},"featured_media":1409466,"template":"","meta":{"_has_post_settings":[]},"schools":[],"areas":[508],"subjects":[416,422],"class_list":["post-1409461","articles","type-articles","status-publish","has-post-thumbnail","hentry","areas-artificial-intelligence","subjects-business-and-finance","subjects-innovation-and-technology"],"custom-fields":{"wpcf-article-leadin":["AI is reshaping credit decisions, but at what cost? Balancing predictive power with transparency is key to ensuring trust in financial systems, write Guillermo de Haro Rodr\u00edguez and Andr\u00e9s Alonso."],"wpcf-article-body":["In the age of artificial intelligence, your chances of getting a loan are increasingly shaped not by a person but by an algorithm. Machine learning (ML) models are now extraordinarily powerful and accurate, able to detect patterns in financial behavior that a traditional loan officer might miss. But they come with a cost: they are remarkably opaque \u2013 and this opacity is more than an inconvenience; it\u2019s a risk because even the model\u2019s own developers may struggle to explain how a specific credit decision was made.\r\n\r\nThese systems can deny someone a loan based on complex patterns buried deep in the data without offering a clear justification. As Cynthia Rudin of Duke University has pointed out, <a href=\"https:\/\/www.nature.com\/articles\/s42256-019-0048-x\" target=\"_blank\" rel=\"noopener\">this lack of transparency is especially dangerous in high-stakes areas<\/a> like credit scoring, where decisions affect people\u2019s access to financial security and future opportunities.\r\n\r\nIndeed, the European Union\u2019s Artificial Intelligence Act has formally classified credit scoring as a \u201chigh-risk\u201d use case, putting pressure on financial institutions to ensure that automated decisions are not only effective but explainable. This raises a fundamental question for the future of AI in finance: How much accuracy (if any) are we willing to trade for transparency?\r\n\r\nMachine learning models like XGBoost have become the go-to tools in credit scoring, thanks to their high accuracy and ability to model complex, non-linear patterns in large, high-dimensional datasets. But their \"black box\" nature makes them difficult to interpret. And that\u2019s a growing concern for both regulators and consumers. Explainability is required not only by regulation but also by consumers seeking transparency in how their financial lives are evaluated.\r\n\r\nReturning to the issue of opacity, this growing concern has led to the development of new techniques aimed at improving model transparency. Many institutions are playing catchup at the moment \u2013 understandable, given the current speed of AI adoption, but the impact can be significant. For example, the Bank of England and the Financial Conduct Authority conducted <a href=\"https:\/\/www.bankofengland.co.uk\/report\/2024\/artificial-intelligence-in-uk-financial-services-2024\" target=\"_blank\" rel=\"noopener\">a survey of AI and machine learning in UK financial services<\/a> and found that while 75% of firms are already using AI, with an additional 10% planning to do so over the next three years, 46% of these reported only a partial understanding of the technologies they use. That gap between usage and understanding highlights the urgent need for more interpretable systems.\r\n\r\nIn response, two different approaches have emerged. First, explainability or post-hoc interpretability techniques which work from the model. For instance, after any model is trained, we may use methods based on the permutation of the features to assess the extent to which they affect our outcome. One of the more promising techniques is Shapley Additive Explanations (SHAP), which draws on game theory to create a set of features' coalitions to rank their local and global importance. These tools are powerful but have their own limitations. They often require high computing power and rely on approximations, meaning that they are missing the statistical rigor of traditional inference. They help open the black box, but they don\u2019t provide a full understanding of it.\r\n\r\nAn alternative approach is to design machine learning models that are interpretable from the start. For instance, using monotonicity constraints as rules embedded into the model that ensure relationships between key variables and predicted outcomes follow consistent, economically sound logic. Typically, in a credit scoring setting, higher income should not lead to higher risk of default. However, free versions of XGBoost models might find correlations, or interactions that imply a non-monotonic (i.e.: constantly increasing or decreasing) relationship between income and probability of default.\r\n\r\nBut this introduces a new challenge, and one that motivated our research: what is the economic and\/ or social cost of this clarity?\r\n\r\nIn a recent empirical study with our colleagues Jose Manuel Carbo of the Banco de Espa\u00f1a and Juan Jos\u00e9 Guill\u00e9n Garc\u00eda of the Universidad Politecnica de Madrid, we <a href=\"https:\/\/cadmus.eui.eu\/server\/api\/core\/bitstreams\/3f2a40e5-0711-4388-acb7-74ae7b328af7\/content\" target=\"_blank\" rel=\"noopener\">compared two versions of an XGBoost model trained on data from LendingClub<\/a>, a major U.S. peer-to-peer lending platform. One version was unconstrained and optimized purely for predictive accuracy. The other applied monotonicity constraints to ensure that the relationships between variables follow financial intuition and supervisory expectations.\r\n\r\nPredictably, the constrained model performed slightly worse in predictive accuracy. On the surface, the difference was in the second decimal. This may seem negligible, but in credit scoring, where misclassifying even a small percentage of loans can lead to substantial financial loss or unfair outcomes, it raises real concerns.\r\n<blockquote>The path forward is not about choosing between innovation and accountability.<\/blockquote>\r\nHowever, we found that the constraints significantly improved the model's explainability without meaningfully disrupting the underlying logic of the predictions. The model became easier to interpret while still relying on the same key variables. In both versions, the top contributing factors (for example, loan term, installment size, income, and FICO score) measured using SHAP, remained largely consistent at global level.\r\n\r\nOn another side, at local level, one of the most striking effects we observed was how the constrained model altered individual credit predictions. Rather than producing scores across a wide spectrum of risk, the model tended to pull estimates toward the mean \u2013 \"mean reversion\". High-risk scores were lowered while low-risk scores pushed higher.\r\n\r\nThis centralizing effect had two notable consequences. First, it reduced granularity. The model became less effective at identifying edge cases, such as those applicants who were truly risky or exceptionally safe. Second, it raised subtle fairness questions. By smoothing out the distribution of risk, the model appears more interpretable but may unintentionally shift scores in ways that disproportionately affect certain demographic and socioeconomic groups more than others.\r\n\r\nFor example, our findings show that some borrowers initially assessed as low risk \u2013 often younger, higher-income individuals \u2013 saw their predicted default probabilities increase. Conversely, some actual defaulters saw their risk scores decrease more sharply than those who repaid their loans. This helps explain the slight dip in overall model accuracy: the constrained version was less likely to flag borrowers at either end of the spectrum.\r\n\r\nThis then raised a second question: even if a model is easier to explain, how do we know those explanations are statistically reliable? To bridge the gap between transparency and rigor, we used a technique called Shapley regressions to test whether the model\u2019s explanations held up under scrutiny.\r\n\r\nThis approach quantifies how much each variable contributes to a prediction, using p-values and confidence intervals \u2013 familiar tools in traditional statistical analysis. The results were promising: the variables most critical to model predictions were also those most statistically significant, and this remained true in both the constrained and unconstrained models. On other words, adding interpretability didn\u2019t weaken the model's data-driven foundation, it clarified it.\r\n\r\nFor financial institutions, that distinction is critical. It allows compliance teams, auditors, and regulators to ask not whether a model works but whether its reasoning is reliable, trustworthy \u2013 an important distinction in environments subject to increasing regulatory scrutiny. Overall, financial institutions must rethink how they evaluate AI model risk. Tools like Shapley regressions can help institutions verify that explainability is evidence-based \u2013 and when paired with \"human-in-the-loop\" oversight, it is more likely that automated decisions align with institutional values and social expectations.\r\n\r\nThe path forward is not about choosing between innovation and accountability. It's about designing systems that optimize both, leveraging the well-known \"multiplicity of good models\" theory of machine learning. It\u2019s possible to improve model transparency and maintain performance \u2013 if institutions rethink how they evaluate risk, measure fairness, and communicate AI-driven decisions.\r\n\r\nInherently interpretable models, supported by statistical validation tools like Shapley regressions, offer a new blueprint for responsible AI adoption, one that reinforces trust with customers, withstand regulatory scrutiny and helps financial institutions understand both the what and the why behind a model\u2019s predictions.\r\n\r\nThe challenge is as cultural as it is technical. Financial organizations and fintech firms must make deliberate choices about how they design and govern their AI systems. That means building teams that combine data science and domain expertise in order to ensure the models used reflect real-world constraints. It also means prioritizing an oversight structure that keeps humans in the loop as a way to reinforce accountability.\r\n\r\nThese choices matter because early design decisions tend to become long-term practices. If AI-driven credit decisions remain opaque and inconsistent, if they are left in the black box, we risk automating inequality and bias. Trust in financial systems will eventually erode. By embracing explainability, it\u2019s possible to develop credit systems that are a benefit to the individuals whose financial futures hang in the balance as well as for the long-term health of the system itself.\r\n\r\n&nbsp;\r\n\r\n\u00a9 IE Insights."],"wpcf-audio-article":["https:\/\/www.ie.edu\/insights\/wp-content\/uploads\/2025\/07\/PlayAI_Rethinking_AI_in_Credit_Decision_Making.mp3"],"wpcf-article-extract":["AI is reshaping credit decisions, but at what cost? Balancing predictive power with transparency is key to ensuring trust in financial systems, write Guillermo de Haro Rodr\u00edguez and Andr\u00e9s Alonso."],"wpcf-article-extract-enable":["1"]},"_links":{"self":[{"href":"https:\/\/www.ie.edu\/insights\/wp-json\/wp\/v2\/articles\/1409461","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.ie.edu\/insights\/wp-json\/wp\/v2\/articles"}],"about":[{"href":"https:\/\/www.ie.edu\/insights\/wp-json\/wp\/v2\/types\/articles"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.ie.edu\/insights\/wp-json\/wp\/v2\/media\/1409466"}],"wp:attachment":[{"href":"https:\/\/www.ie.edu\/insights\/wp-json\/wp\/v2\/media?parent=1409461"}],"wp:term":[{"taxonomy":"schools","embeddable":true,"href":"https:\/\/www.ie.edu\/insights\/wp-json\/wp\/v2\/schools?post=1409461"},{"taxonomy":"areas","embeddable":true,"href":"https:\/\/www.ie.edu\/insights\/wp-json\/wp\/v2\/areas?post=1409461"},{"taxonomy":"subjects","embeddable":true,"href":"https:\/\/www.ie.edu\/insights\/wp-json\/wp\/v2\/subjects?post=1409461"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}