Machine learning interpretability for a stress scenario generation in credit scoring based on counterfactuals

Andreas C. Bueff*, Mateusz Cytrynski, Raffaella Calabrese, Matthew Jones, John Roberts, Jonathon Moore, Iain Brown

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract / Description of output

To boost the application of machine learning (ML) techniques for credit scoring models, the blackbox problem should be addressed. The primary aim of this paper is to propose a measure based on counterfactuals to evaluate the interpretability of a ML credit scoring technique. Counterfactuals assist with understanding the model with regard to the classification decision boundaries and evaluate model robustness. The second contribution is the development of a data perturbation technique to generate a stress scenario.

We apply these two proposals to a dataset on UK unsecured personal loans to compare logistic regression and stochastic gradient boosting (SBG). We show that training a blackbox model (SGB) as conditioned on our data perturbation technique can provide insight into model performance under stressed scenarios. The empirical results show that our interpretability measure is able to capture the classification decision boundary, unlike AUC and the classification accuracy widely used in the banking sector.
Original languageEnglish
Article number117271
JournalExpert Systems with Applications
Volume202
Early online date26 Apr 2022
DOIs
Publication statusPublished - 15 Sept 2022

Keywords / Materials (for Non-textual outputs)

  • OR in banking
  • interpretable ML
  • credit scoring
  • stress scenario

Fingerprint

Dive into the research topics of 'Machine learning interpretability for a stress scenario generation in credit scoring based on counterfactuals'. Together they form a unique fingerprint.

Cite this