School of Computer Science, McGill University, Montreal, QC H3A 0E9, Canada.
Proc Natl Acad Sci U S A. 2023 Mar 7;120(10):e2209384120. doi: 10.1073/pnas.2209384120. Epub 2023 Feb 27.
The machine learning (ML) research community has landed on automated hate speech detection as the vital tool in the mitigation of bad behavior online. However, it is not clear that this is a widely supported view outside of the ML world. Such a disconnect can have implications for whether automated detection tools are accepted or adopted. Here we lend insight into how other key stakeholders understand the challenge of addressing hate speech and the role automated detection plays in solving it. To do so, we develop and apply a structured approach to dissecting the discourses used by online platform companies, governments, and not-for-profit organizations when discussing hate speech. We find that, where hate speech mitigation is concerned, there is a profound disconnect between the computer science research community and other stakeholder groups-which puts progress on this important problem at serious risk. We identify urgent steps that need to be taken to incorporate computational researchers into a single, coherent, multistakeholder community that is working towards civil discourse online.
机器学习(ML)研究社区已经将自动仇恨言论检测作为减轻网络不良行为的重要工具。然而,在 ML 领域之外,这是否是一个广泛支持的观点尚不清楚。这种脱节可能会影响到自动化检测工具是否被接受或采用。在这里,我们深入了解其他主要利益相关者如何理解解决仇恨言论的挑战,以及自动化检测在解决该问题中所扮演的角色。为此,我们开发并应用了一种结构化的方法来剖析在线平台公司、政府和非营利组织在讨论仇恨言论时使用的话语。我们发现,在减少仇恨言论方面,计算机科学研究社区与其他利益相关群体之间存在着深刻的脱节,这使得这一重要问题的进展面临严重风险。我们确定了需要采取的紧急步骤,将计算研究人员纳入一个单一的、连贯的、多利益相关者社区,共同努力实现在线文明对话。