Abstract / Description of output
Gender is prohibited from use in decision-making in many countries. This does not necessarily benefit females in situations of automated algorithmic decisions, e.g. when a credit scoring model is used as a decision tool for loan granting. By analysing a unique proprietary dataset on car loans from an EU bank, this paper shows that Gender as a variable in a credit scoring model is statistically significant. Its removal does not alter the predictive accuracy of the model, yet the proportions of accepted women/men depend on whether Gender is included. The paper explores the association between predictors in the model with Gender, to demonstrate the omitted variable bias and how other variables proxy for Gender. It points to inconsistencies of the existing regulations in the context of automated decision-making.
Original language | English |
---|---|
Pages (from-to) | 1287-1311 |
Journal | Journal of the Royal Statistical Society: Statistics in Society Series A |
Volume | 182 |
Issue number | 4 |
Early online date | 1 Aug 2019 |
DOIs | |
Publication status | Published - 25 Oct 2019 |
Keywords / Materials (for Non-textual outputs)
- credit scoring
- gender
- algorithmic decision-making
- statistical discrimination
Fingerprint
Dive into the research topics of 'The law of equal opportunities or unintended consequences? The impact of unisex risk assessment in consumer credit'. Together they form a unique fingerprint.Profiles
-
Galina Andreeva
- Business School - Personal Chair of Societal Aspects of Credit
- Management Science and Business Economics
- Credit Research Centre
- Management Science
- Edinburgh Centre for Financial Innovations
Person: Academic: Research Active