Program in Audiology, Department of Otolaryngology Head and Neck Surgery and Communicative Disorders, School of Medicine, University of Louisville, KY.
Department of Communication Sciences and Disorders, University of Utah, Salt Lake City.
J Speech Lang Hear Res. 2019 Nov 22;62(11):4015-4029. doi: 10.1044/2019_JSLHR-S-19-0144. Epub 2019 Oct 25.
Purpose Emotion classification for auditory stimuli typically employs 1 of 2 approaches (discrete categories or emotional dimensions). This work presents a new emotional speech set, compares these 2 classification methods for emotional speech stimuli, and emphasizes the need to consider the entire communication model (i.e., the talker, message, and listener) when studying auditory emotion portrayal and perception. Method Emotional speech from male and female talkers was evaluated using both categorical and dimensional rating methods. Ten young adult listeners (ages 19-28 years) evaluated stimuli recorded in 4 emotional speaking styles (Angry, Calm, Happy, and Sad). Talker and listener factors were examined for potential influences on emotional ratings using categorical and dimensional rating methods. Listeners rated stimuli by selecting an emotion category, rating the activation and pleasantness, and indicating goodness of category fit. Results Discrete ratings were generally consistent with dimensional ratings for speech, with accuracy for emotion recognition well above chance. As stimuli approached dimensional extremes of activation and pleasantness, listeners were more confident in their category selection, indicative of a hybrid approach to emotion classification. Female talkers were rated as more activated than male talkers, and female listeners gave higher ratings of activation compared to male listeners, confirming gender differences in emotion perception. Conclusion A hybrid model for auditory emotion classification is supported by the data. Talker and listener factors, such as gender, were found to impact the ratings of emotional speech and must be considered alongside stimulus factors in the design of future studies of emotion.
目的 听觉刺激的情感分类通常采用以下两种方法之一(离散类别或情感维度)。本研究提出了一种新的情感语音集,比较了这两种情感语音刺激的分类方法,并强调在研究听觉情感表达和感知时,需要考虑整个通信模型(即说话者、信息和听众)。
方法 使用分类和维度评分方法评估来自男性和女性说话者的情感语音。10 位年轻成年听众(19-28 岁)评估了以 4 种情感说话风格(愤怒、平静、快乐和悲伤)录制的刺激。使用分类和维度评分方法检查说话者和听众因素对情感评分的潜在影响。听众通过选择情感类别、评定激活度和愉悦度以及指示类别拟合度来对刺激进行评分。
结果 离散评分通常与语音的维度评分一致,情感识别的准确性远高于随机水平。随着刺激接近激活度和愉悦度的维度极端,听众对其类别选择更有信心,这表明情感分类采用了混合方法。女性说话者的活跃度评分高于男性说话者,而女性听众的活跃度评分高于男性听众,这证实了情感感知方面的性别差异。
结论 数据支持听觉情感分类的混合模型。说话者和听众因素(如性别)会影响情感语音的评分,在未来的情感研究设计中,必须与刺激因素一起考虑这些因素。