School of Public Health, The University of Texas Health Science Center at Houston, Houston, TX, United States.
Department of Computer Science & Engineering, Texas A&M University, College Station, TX, United States.
J Biomed Inform. 2023 Jul;143:104399. doi: 10.1016/j.jbi.2023.104399. Epub 2023 May 19.
The emphasis on fairness in predictive healthcare modeling has increased in popularity as an approach for overcoming biases in automated decision-making systems. The aim is to guarantee that sensitive characteristics like gender, race, and ethnicity do not influence prediction outputs. Numerous algorithmic strategies have been proposed to reduce bias in prediction results, mitigate prejudice toward minority groups and promote prediction fairness. The goal of these strategies is to ensure that model prediction performance does not exhibit significant disparity among sensitive groups. In this study, we propose a novel fairness-achieving scheme based on multitask learning, which fundamentally differs from conventional fairness-achieving techniques, including altering data distributions and constraint optimization through regularizing fairness metrics or tampering with prediction outcomes. By dividing predictions on different sub-populations into separate tasks, we view the fairness problem as a task-balancing problem. To ensure fairness during the model-training process, we suggest a novel dynamic re-weighting approach. Fairness is achieved by dynamically modifying the gradients of various prediction tasks during neural network back-propagation, and this novel technique applies to a wide range of fairness criteria. We conduct tests on a real-world use case to predict sepsis patients' mortality risk. Our approach satisfies that it can reduce the disparity between subgroups by 98% while only losing less than 4% of prediction accuracy.
强调公平性在预测医疗建模中已经成为一种克服自动化决策系统偏差的流行方法。其目的是确保像性别、种族和民族这样的敏感特征不会影响预测输出。已经提出了许多算法策略来减少预测结果中的偏差,减轻对少数群体的偏见,并促进预测公平性。这些策略的目的是确保模型的预测性能在敏感群体之间没有显著差异。在这项研究中,我们提出了一种基于多任务学习的新的公平实现方案,它与传统的公平实现技术有根本的不同,包括通过正则化公平度量或篡改预测结果来改变数据分布和约束优化。通过将不同子群体的预测分为单独的任务,我们将公平问题视为任务平衡问题。为了在模型训练过程中实现公平性,我们提出了一种新的动态重新加权方法。通过在神经网络反向传播过程中动态修改各种预测任务的梯度来实现公平性,这种新技术适用于广泛的公平性标准。我们在一个真实的用例中进行了测试,以预测败血症患者的死亡率风险。我们的方法可以将子组之间的差异减少 98%,而只损失不到 4%的预测准确性。