Biradar Shankar, Saumya Sunil, Chauhan Arun
Department of Computer Science and Engineering, Indian Institute of Information Technology, Dharwad, Karnataka India.
Department of Computer Science and Engineering, Graphic Era University, Dehradun, Uttarakhand India.
Soc Netw Anal Min. 2022;12(1):87. doi: 10.1007/s13278-022-00920-w. Epub 2022 Jul 24.
Many people have begun to use social media platforms due to the increased use of the Internet over the previous decade. It has a lot of benefits, but it also comes with a lot of risks and drawbacks, such as Hate speech. People in multilingual societies, such as India, frequently mix their native language with English while speaking, so detecting hate content in such bilingual code-mixed data has drawn the larger interest of the research community. The majority of previous work focuses on high-resource language such as English, but very few researchers have concentrated on the mixed bilingual data like Hinglish. In this study, we investigated the performance of transformer models like IndicBERT and multilingual Bidirectional Encoder Representation(mBERT), as well as transfer learning from pre-trained language models like ULMFiT and Bidirectional encoder Representation(BERT), to find hateful content in Hinglish. Also, Transformer-based Interpreter and Feature extraction model on Deep Neural Network (TIF-DNN), is proposed in this work. The experimental results found that our proposed model outperforms existing state-of-art methods for Hate speech identification in Hinglish language with an accuracy of 73%.
在过去十年中,由于互联网使用的增加,许多人开始使用社交媒体平台。它有很多好处,但也伴随着很多风险和缺点,比如仇恨言论。在印度这样的多语言社会中,人们在说话时经常将母语与英语混合使用,因此在这种双语代码混合数据中检测仇恨内容引起了研究界的更大兴趣。以前的大多数工作都集中在英语等资源丰富的语言上,但很少有研究人员专注于像印地语-英语这样的混合双语数据。在本研究中,我们研究了像IndicBERT和多语言双向编码器表示(mBERT)这样的变压器模型的性能,以及从像ULMFiT和双向编码器表示(BERT)这样的预训练语言模型进行迁移学习,以在印地语-英语中找到仇恨内容。此外,本文还提出了基于变压器的深度神经网络解释器和特征提取模型(TIF-DNN)。实验结果发现,我们提出的模型在印地语-英语仇恨言论识别方面优于现有的最先进方法,准确率达到73%。