Szepannek Gero, Lübke Karsten
Institute of Applied Computer Science, Stralsund University of Applied Sciences, Stralsund, Germany.
Institute for Empirical Research and Statistics, FOM University of Applied Sciences, Dortmund, Germany.
Front Artif Intell. 2021 Oct 14;4:681915. doi: 10.3389/frai.2021.681915. eCollection 2021.
Algorithmic scoring methods are widely used in the finance industry for several decades in order to prevent risk and to automate and optimize decisions. Regulatory requirements as given by the Basel Committee on Banking Supervision (BCBS) or the EU data protection regulations have led to an increasing interest and research activity on understanding black box machine learning models by means of explainable machine learning. Even though this is a step into a right direction, such methods are not able to guarantee for a fair scoring as machine learning models are not necessarily unbiased and may discriminate with respect to certain subpopulations such as a particular race, gender, or sexual orientation-even if the variable itself is not used for modeling. This is also true for white box methods like logistic regression. In this study, a framework is presented that allows analyzing and developing models with regard to fairness. The proposed methodology is based on techniques of causal inference and some of the methods can be linked to methods from explainable machine learning. A definition of counterfactual fairness is given together with an algorithm that results in a fair scoring model. The concepts are illustrated by means of a transparent simulation and a popular real-world example, the German Credit data using traditional scorecard models based on logistic regression and weight of evidence variable pre-transform. In contrast to previous studies in the field for our study, a corrected version of the data is presented and used. With the help of the simulation, the trade-off between fairness and predictive accuracy is analyzed. The results indicate that it is possible to remove unfairness without a strong performance decrease unless the correlation of the discriminative attributes on the other predictor variables in the model is not too strong. In addition, the challenge in explaining the resulting scoring model and the associated fairness implications to users is discussed.
几十年来,算法评分方法在金融行业中被广泛使用,以防范风险并实现决策的自动化和优化。巴塞尔银行监管委员会(BCBS)给出的监管要求或欧盟数据保护法规,引发了人们对通过可解释机器学习来理解黑箱机器学习模型的兴趣日益浓厚,并推动了相关研究活动。尽管这是朝着正确方向迈出的一步,但此类方法无法保证公平评分,因为机器学习模型不一定是无偏的,可能会对某些亚群体(如特定种族、性别或性取向)产生歧视——即使变量本身未用于建模。逻辑回归等白盒方法也是如此。在本研究中,提出了一个框架,该框架允许对模型的公平性进行分析和开发。所提出的方法基于因果推断技术,其中一些方法可以与可解释机器学习中的方法相联系。给出了反事实公平性的定义以及一个能产生公平评分模型的算法。通过一个透明模拟和一个流行的实际例子(即使用基于逻辑回归和证据权重变量预变换的传统计分卡模型的德国信贷数据)对这些概念进行了说明。与该领域之前的研究不同,本研究给出并使用了数据的校正版本。借助模拟,分析了公平性与预测准确性之间的权衡。结果表明,除非模型中判别属性与其他预测变量之间的相关性不太强,否则有可能在不显著降低性能的情况下消除不公平性。此外,还讨论了向用户解释所得评分模型及其相关公平性影响方面的挑战。