Balakrishnan Vimala, Shi Zhongliang, Law Chuan Liang, Lim Regine, Teh Lee Leng, Fan Yue
Faculty of Computer Science and Information Technology, Universiti Malaya, 50603 Kuala Lumpur, Malaysia.
Malayan Banking Berhad, 50050 Kuala Lumpur, Malaysia.
J Supercomput. 2022;78(5):7206-7226. doi: 10.1007/s11227-021-04169-6. Epub 2021 Nov 5.
We present a benchmark comparison of several deep learning models including Convolutional Neural Networks, Recurrent Neural Network and Bi-directional Long Short Term Memory, assessed based on various word embedding approaches, including the Bi-directional Encoder Representations from Transformers (BERT) and its variants, FastText and Word2Vec. Data augmentation was administered using the Easy Data Augmentation approach resulting in two datasets (original versus augmented). All the models were assessed in two setups, namely 5-class versus 3-class (i.e., compressed version). Findings show the best prediction models were Neural Network-based using Word2Vec, with CNN-RNN-Bi-LSTM producing the highest accuracy (96%) and -score (91.1%). Individually, RNN was the best model with an accuracy of 87.5% and -score of 83.5%, while RoBERTa had the best -score of 73.1%. The study shows that deep learning is better for analyzing the sentiments within the text compared to supervised machine learning and provides a direction for future work and research.
我们展示了几种深度学习模型的基准比较,包括卷积神经网络、循环神经网络和双向长短期记忆网络,这些模型基于各种词嵌入方法进行评估,包括来自变换器的双向编码器表示(BERT)及其变体、FastText和Word2Vec。使用简易数据增强方法进行数据增强,从而得到两个数据集(原始数据集与增强数据集)。所有模型在两种设置下进行评估,即5类与3类(即压缩版本)。研究结果表明,使用Word2Vec的基于神经网络的预测模型最佳,其中CNN-RNN-Bi-LSTM的准确率最高(96%),F1分数最高(91.1%)。单独来看,RNN是最佳模型,准确率为87.5%,F1分数为83.5%,而RoBERTa的F1分数最高,为73.1%。该研究表明,与监督式机器学习相比,深度学习在分析文本中的情感方面表现更佳,并为未来的工作和研究提供了方向。