Xinjiang Key Laboratory of Signal Detection and Processing, College of Information Science and Engineering, Xinjiang University, Urumqi 830046, China.
College of Information Science and Engineering, Xinjiang University, Urumqi 830046, China.
Sensors (Basel). 2023 Mar 1;23(5):2679. doi: 10.3390/s23052679.
Multimodal sentiment analysis has gained popularity as a research field for its ability to predict users' emotional tendencies more comprehensively. The data fusion module is a critical component of multimodal sentiment analysis, as it allows for integrating information from multiple modalities. However, it is challenging to combine modalities and remove redundant information effectively. In our research, we address these challenges by proposing a multimodal sentiment analysis model based on supervised contrastive learning, which leads to more effective data representation and richer multimodal features. Specifically, we introduce the MLFC module, which utilizes a convolutional neural network (CNN) and Transformer to solve the redundancy problem of each modal feature and reduce irrelevant information. Moreover, our model employs supervised contrastive learning to enhance its ability to learn standard sentiment features from data. We evaluate our model on three widely-used datasets, namely MVSA-single, MVSA-multiple, and HFM, demonstrating that our model outperforms the state-of-the-art model. Finally, we conduct ablation experiments to validate the efficacy of our proposed method.
多模态情感分析因其能够更全面地预测用户的情感倾向而成为一个热门的研究领域。数据融合模块是多模态情感分析的关键组成部分,因为它允许整合来自多个模态的信息。然而,有效地结合模态并去除冗余信息是具有挑战性的。在我们的研究中,我们通过提出一种基于监督对比学习的多模态情感分析模型来解决这些挑战,该模型导致更有效的数据表示和更丰富的多模态特征。具体来说,我们引入了 MLFC 模块,该模块利用卷积神经网络 (CNN) 和 Transformer 来解决每个模态特征的冗余问题,并减少无关信息。此外,我们的模型采用监督对比学习来增强其从数据中学习标准情感特征的能力。我们在三个广泛使用的数据集上评估了我们的模型,即 MVSA-single、MVSA-multiple 和 HFM,结果表明我们的模型优于最先进的模型。最后,我们进行了消融实验来验证我们提出的方法的有效性。