Suppr超能文献

从语言语料库中自动推导出来的语义包含类人偏见。

Semantics derived automatically from language corpora contain human-like biases.

机构信息

Center for Information Technology Policy, Princeton University, Princeton, NJ, USA.

Department of Computer Science, University of Bath, Bath BA2 7AY, UK.

出版信息

Science. 2017 Apr 14;356(6334):183-186. doi: 10.1126/science.aal4230.

Abstract

Machine learning is a means to derive artificial intelligence by discovering patterns in existing data. Here, we show that applying machine learning to ordinary human language results in human-like semantic biases. We replicated a spectrum of known biases, as measured by the Implicit Association Test, using a widely used, purely statistical machine-learning model trained on a standard corpus of text from the World Wide Web. Our results indicate that text corpora contain recoverable and accurate imprints of our historic biases, whether morally neutral as toward insects or flowers, problematic as toward race or gender, or even simply veridical, reflecting the status quo distribution of gender with respect to careers or first names. Our methods hold promise for identifying and addressing sources of bias in culture, including technology.

摘要

机器学习是通过发现现有数据中的模式来获得人工智能的一种手段。在这里,我们表明,将机器学习应用于普通人类语言会导致类似人类的语义偏见。我们使用广泛使用的、仅基于统计的机器学习模型,该模型基于来自万维网的标准文本语料库进行训练,复制了一系列已知的偏见,这些偏见是通过内隐联想测试来衡量的。我们的结果表明,文本语料库包含可恢复和准确的历史偏见印记,无论是对昆虫或花朵等道德中立的偏见,还是对种族或性别等有问题的偏见,甚至是简单的真实性偏见,反映了性别在职业或名字方面的现状分布。我们的方法有望识别和解决文化中的偏见来源,包括技术。

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验