Laaksonen Salla-Maaria, Haapoja Jesse, Kinnunen Teemu, Nelimarkka Matti, Pöyhtäri Reeta
Centre for Consumer Society Research, University of Helsinki, Helsinki, Finland.
Department of Computer Science, Aalto University, Espoo, Finland.
Front Big Data. 2020 Feb 5;3:3. doi: 10.3389/fdata.2020.00003. eCollection 2020.
Hate speech has been identified as a pressing problem in society and several automated approaches have been designed to detect and prevent it. This paper reports and reflects upon an action research setting consisting of multi-organizational collaboration conducted during Finnish municipal elections in 2017, wherein a technical infrastructure was designed to automatically monitor candidates' social media updates for hate speech. The setting allowed us to engage in a 2-fold investigation. First, the collaboration offered a unique view for exploring how hate speech emerges as a technical problem. The project developed an adequately well-working algorithmic solution using supervised machine learning. We tested the performance of various feature extraction and machine learning methods and ended up using a combination of Bag-of-Words feature extraction with Support-Vector Machines. However, an automated approach required heavy simplification, such as using rudimentary scales for classifying hate speech and a reliance on word-based approaches, while in reality hate speech is a linguistic and social phenomenon with various tones and forms. Second, the action-research-oriented setting allowed us to observe affective responses, such as the hopes, dreams, and fears related to machine learning technology. Based on participatory observations, project artifacts and documents, interviews with project participants, and online reactions to the detection project, we identified participants' aspirations for effective automation as well as the level of neutrality and objectivity introduced by an algorithmic system. However, the participants expressed more critical views toward the system after the monitoring process. Our findings highlight how the powerful expectations related to technology can easily end up dominating a project dealing with a contested, topical social issue. We conclude by discussing the problematic aspects of datafying hate and suggesting some practical implications for hate speech recognition.
仇恨言论已被视为社会中的一个紧迫问题,并且已经设计了几种自动化方法来检测和预防它。本文报告并反思了一个行动研究场景,该场景包括在2017年芬兰市政选举期间进行的多组织合作,其中设计了一个技术基础设施来自动监测候选人社交媒体更新中的仇恨言论。这个场景使我们能够进行两方面的调查。首先,这种合作提供了一个独特的视角来探索仇恨言论如何作为一个技术问题出现。该项目使用监督式机器学习开发了一个运行良好的算法解决方案。我们测试了各种特征提取和机器学习方法的性能,最终使用了词袋特征提取与支持向量机的组合。然而,自动化方法需要大量简化,例如使用基本量表来分类仇恨言论以及依赖基于单词的方法,而实际上仇恨言论是一种具有各种语气和形式的语言和社会现象。其次,以行动研究为导向的场景使我们能够观察情感反应,例如与机器学习技术相关的希望、梦想和恐惧。基于参与式观察、项目工件和文档、对项目参与者的访谈以及对检测项目的在线反应,我们确定了参与者对有效自动化的期望以及算法系统引入的中立性和客观性水平。然而,在监测过程之后,参与者对该系统表达了更批判性的观点。我们的研究结果强调了与技术相关的强大期望如何轻易地最终主导一个处理有争议的热点社会问题的项目。我们通过讨论将仇恨数据化的问题方面并提出一些仇恨言论识别的实际意义来得出结论。