Lin Chang-Hong, Liao Wei-Kai, Hsieh Wen-Chi, Liao Wei-Jiun, Wang Jia-Ching
Department of Computer Science and Information Engineering, National Central University, Taiwan.
ScientificWorldJournal. 2014;2014:757121. doi: 10.1155/2014/757121. Epub 2014 May 20.
The investigations of emotional speech identification can be divided into two main parts, features and classifiers. In this paper, how to extract an effective speech feature set for the emotional speech identification is addressed. In our speech feature set, we use not only statistical analysis of frame-based acoustical features, but also the approximated speech feature contours, which are obtained by extracting extremely low frequency components to speech feature contours. Furthermore, principal component analysis (PCA) is applied to the approximated speech feature contours so that an efficient representation of approximated contours can be derived. The proposed speech feature set is fed into support vector machines (SVMs) to perform multiclass emotion identification. The experimental results demonstrate the performance of the proposed system with 82.26% identification rate.
对情感语音识别的研究可分为两个主要部分,即特征和分类器。本文探讨了如何为情感语音识别提取有效的语音特征集。在我们的语音特征集中,我们不仅使用基于帧的声学特征的统计分析,还使用近似语音特征轮廓,这些轮廓是通过提取语音特征轮廓的极低频成分获得的。此外,主成分分析(PCA)应用于近似语音特征轮廓,以便能够导出近似轮廓的有效表示。所提出的语音特征集被输入到支持向量机(SVM)中以执行多类情感识别。实验结果表明,所提出系统的识别率为82.26%,证明了其性能。