Stinson Catherine
Philosophy Department and School of Computing Kingston, Queen's University at Kingston, Kingston, ON Canada.
AI Ethics. 2022;2(4):763-770. doi: 10.1007/s43681-022-00136-w. Epub 2022 Jan 31.
When Artificial Intelligence (AI) is applied in decision-making that affects people's lives, it is now well established that the outcomes can be biased or discriminatory. The question of whether algorithms themselves can be among the sources of bias has been the subject of recent debate among Artificial Intelligence researchers, and scholars who study the social impact of technology. There has been a tendency to focus on examples, where the data set used to train the AI is biased, and denial on the part of some researchers that algorithms can also be biased. Here we illustrate the point that algorithms themselves can be the source of bias with the example of collaborative filtering algorithms for recommendation and search. These algorithms are known to suffer from cold-start, popularity, and homogenizing biases, among others. While these are typically described as statistical biases rather than biases of moral import; in this paper we show that these statistical biases can lead directly to discriminatory outcomes. The intuitive idea is that data points on the margins of distributions of human data tend to correspond to marginalized people. The statistical biases described here have the effect of further marginalizing the already marginal. Biased algorithms for applications such as media recommendations can have significant impact on individuals' and communities' access to information and culturally-relevant resources. This source of bias warrants serious attention given the ubiquity of algorithmic decision-making.
当人工智能(AI)应用于影响人们生活的决策时,现在已经明确其结果可能存在偏差或歧视性。算法本身是否可能是偏差来源之一,一直是人工智能研究人员以及研究技术社会影响的学者近期争论的话题。人们倾向于关注用于训练AI的数据集存在偏差的例子,并且一些研究人员否认算法也会有偏差。在此,我们以推荐和搜索的协同过滤算法为例,来说明算法本身可能是偏差来源这一观点。众所周知,这些算法存在冷启动、流行度和同质化偏差等问题。虽然这些通常被描述为统计偏差而非具有道德意义的偏差;但在本文中我们表明,这些统计偏差可能直接导致歧视性结果。直观的想法是,人类数据分布边缘的数据点往往对应于被边缘化的人群。这里描述的统计偏差会进一步边缘化那些已经处于边缘地位的人。诸如媒体推荐等应用中的偏差算法,可能对个人和社区获取信息及文化相关资源产生重大影响。鉴于算法决策的普遍性,这种偏差来源值得认真关注。