IEEE Trans Vis Comput Graph. 2024 Jan;30(1):327-337. doi: 10.1109/TVCG.2023.3327192. Epub 2023 Dec 27.
Machine learning technology has become ubiquitous, but, unfortunately, often exhibits bias. As a consequence, disparate stakeholders need to interact with and make informed decisions about using machine learning models in everyday systems. Visualization technology can support stakeholders in understanding and evaluating trade-offs between, for example, accuracy and fairness of models. This paper aims to empirically answer "Can visualization design choices affect a stakeholder's perception of model bias, trust in a model, and willingness to adopt a model?" Through a series of controlled, crowd-sourced experiments with more than 1,500 participants, we identify a set of strategies people follow in deciding which models to trust. Our results show that men and women prioritize fairness and performance differently and that visual design choices significantly affect that prioritization. For example, women trust fairer models more often than men do, participants value fairness more when it is explained using text than as a bar chart, and being explicitly told a model is biased has a bigger impact than showing past biased performance. We test the generalizability of our results by comparing the effect of multiple textual and visual design choices and offer potential explanations of the cognitive mechanisms behind the difference in fairness perception and trust. Our research guides design considerations to support future work developing visualization systems for machine learning.
机器学习技术已经无处不在,但不幸的是,它常常存在偏见。因此,不同的利益相关者需要互动,并就如何在日常系统中使用机器学习模型做出明智的决策。可视化技术可以支持利益相关者理解和评估模型的准确性和公平性等方面之间的权衡。本文旨在通过一系列有超过 1500 名参与者参与的受控、众包实验,从经验上回答“可视化设计选择是否会影响利益相关者对模型偏差的感知、对模型的信任以及采用模型的意愿?”我们确定了人们在决定信任哪些模型时遵循的一系列策略。我们的研究结果表明,男性和女性对公平性和性能的重视程度不同,并且视觉设计选择会显著影响这种优先级排序。例如,女性比男性更信任公平性更高的模型,参与者更重视通过文本而不是条形图来解释的公平性,并且明确告知模型存在偏差比显示过去的偏差性能产生更大的影响。我们通过比较多种文本和视觉设计选择的效果来测试我们结果的普遍性,并为公平感知和信任背后的认知机制差异提供潜在的解释。我们的研究为开发机器学习可视化系统的未来工作提供了设计考虑因素的指导。