Shakil Md Shahid Ahammed, Farid Fahmid Al, Podder Nitun Kumar, Iqbal S M Hasan Sazzad, Miah Abu Saleh Musa, Rahim Md Abdur, Karim Hezerul Abdul
Department of Computer Science and Engineering, Pabna University of Science and Technology, Pabna 6600, Bangladesh.
Centre for Image and Vision Computing (CIVC), COE for Artificial Intelligence, Faculty of Artificial Intelligence and Engineering (FAIE), Multimedia University, Cyberjaya 63100, Selangor, Malaysia.
J Imaging. 2025 Aug 14;11(8):273. doi: 10.3390/jimaging11080273.
Emotion recognition in speech is essential for enhancing human-computer interaction (HCI) systems. Despite progress in Bangla speech emotion recognition, challenges remain, including low accuracy, speaker dependency, and poor generalization across emotional expressions. Previous approaches often rely on traditional machine learning or basic deep learning models, struggling with robustness and accuracy in noisy or varied data. In this study, we propose a novel multi-stream deep learning feature fusion approach for Bangla speech emotion recognition, addressing the limitations of existing methods. Our approach begins with various data augmentation techniques applied to the training dataset, enhancing the model's robustness and generalization. We then extract a comprehensive set of handcrafted features, including Zero-Crossing Rate (ZCR), chromagram, spectral centroid, spectral roll-off, spectral contrast, spectral flatness, Mel-Frequency Cepstral Coefficients (MFCCs), Root Mean Square (RMS) energy, and Mel-spectrogram. Although these features are used as 1D numerical vectors, some of them are computed from time-frequency representations (e.g., chromagram, Mel-spectrogram) that can themselves be depicted as images, which is conceptually close to imaging-based analysis. These features capture key characteristics of the speech signal, providing valuable insights into the emotional content. Sequentially, we utilize a multi-stream deep learning architecture to automatically learn complex, hierarchical representations of the speech signal. This architecture consists of three distinct streams: the first stream uses 1D convolutional neural networks (1D CNNs), the second integrates 1D CNN with Long Short-Term Memory (LSTM), and the third combines 1D CNNs with bidirectional LSTM (Bi-LSTM). These models capture intricate emotional nuances that handcrafted features alone may not fully represent. For each of these models, we generate predicted scores and then employ ensemble learning with a soft voting technique to produce the final prediction. This fusion of handcrafted features, deep learning-derived features, and ensemble voting enhances the accuracy and robustness of emotion identification across multiple datasets. Our method demonstrates the effectiveness of combining various learning models to improve emotion recognition in Bangla speech, providing a more comprehensive solution compared with existing methods. We utilize three primary datasets-SUBESCO, BanglaSER, and a merged version of both-as well as two external datasets, RAVDESS and EMODB, to assess the performance of our models. Our method achieves impressive results with accuracies of 92.90%, 85.20%, 90.63%, 67.71%, and 69.25% for the SUBESCO, BanglaSER, merged SUBESCO and BanglaSER, RAVDESS, and EMODB datasets, respectively. These results demonstrate the effectiveness of combining handcrafted features with deep learning-based features through ensemble learning for robust emotion recognition in Bangla speech.
语音中的情感识别对于增强人机交互(HCI)系统至关重要。尽管孟加拉语语音情感识别取得了进展,但挑战依然存在,包括准确率低、依赖说话者以及跨情感表达的泛化能力差。以前的方法通常依赖传统机器学习或基本深度学习模型,在处理噪声或多样数据时难以保证鲁棒性和准确性。在本研究中,我们提出了一种用于孟加拉语语音情感识别的新型多流深度学习特征融合方法,以解决现有方法的局限性。我们的方法首先对训练数据集应用各种数据增强技术,以提高模型的鲁棒性和泛化能力。然后,我们提取一组全面的手工特征,包括过零率(ZCR)、色度图、谱质心、谱滚降、谱对比度、谱平坦度、梅尔频率倒谱系数(MFCC)、均方根(RMS)能量和梅尔频谱图。尽管这些特征被用作一维数值向量,但其中一些是从时频表示(例如,色度图、梅尔频谱图)计算得出的,这些时频表示本身可以描绘为图像,这在概念上与基于成像的分析相近。这些特征捕获了语音信号的关键特征,为情感内容提供了有价值的见解。接下来,我们利用多流深度学习架构自动学习语音信号的复杂分层表示。该架构由三个不同的流组成:第一个流使用一维卷积神经网络(1D CNN),第二个流将一维卷积神经网络与长短期记忆(LSTM)集成,第三个流将一维卷积神经网络与双向长短期记忆(Bi-LSTM)相结合。这些模型捕获了单独的手工特征可能无法完全表示的复杂情感细微差别。对于这些模型中的每一个,我们生成预测分数,然后采用带有软投票技术的集成学习来产生最终预测。手工特征、深度学习衍生特征和集成投票的这种融合提高了跨多个数据集的情感识别的准确性和鲁棒性。我们的方法证明了结合各种学习模型以改善孟加拉语语音情感识别的有效性,与现有方法相比提供了更全面的解决方案。我们使用三个主要数据集——SUBESCO、BanglaSER 以及两者的合并版本——以及两个外部数据集 RAVDESS 和 EMODB 来评估我们模型的性能。我们的方法分别在 SUBESCO、BanglaSER、合并后的 SUBESCO 和 BanglaSER、RAVDESS 和 EMODB 数据集上取得了令人印象深刻的结果,准确率分别为 92.90%、85.20%、90.63%、67.71%和 69.25%。这些结果证明了通过集成学习将手工特征与基于深度学习的特征相结合以实现孟加拉语语音鲁棒情感识别的有效性。