• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

共享声学编码是音乐和言语中情感交流的基础——来自深度迁移学习的证据。

Shared acoustic codes underlie emotional communication in music and speech-Evidence from deep transfer learning.

作者信息

Coutinho Eduardo, Schuller Björn

机构信息

Department of Music, University of Liverpool, Liverpool, United Kingdom.

Department of Computing, Imperial College London, London, United Kingdom.

出版信息

PLoS One. 2017 Jun 28;12(6):e0179289. doi: 10.1371/journal.pone.0179289. eCollection 2017.

DOI:10.1371/journal.pone.0179289
PMID:28658285
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC5489171/
Abstract

Music and speech exhibit striking similarities in the communication of emotions in the acoustic domain, in such a way that the communication of specific emotions is achieved, at least to a certain extent, by means of shared acoustic patterns. From an Affective Sciences points of view, determining the degree of overlap between both domains is fundamental to understand the shared mechanisms underlying such phenomenon. From a Machine learning perspective, the overlap between acoustic codes for emotional expression in music and speech opens new possibilities to enlarge the amount of data available to develop music and speech emotion recognition systems. In this article, we investigate time-continuous predictions of emotion (Arousal and Valence) in music and speech, and the Transfer Learning between these domains. We establish a comparative framework including intra- (i.e., models trained and tested on the same modality, either music or speech) and cross-domain experiments (i.e., models trained in one modality and tested on the other). In the cross-domain context, we evaluated two strategies-the direct transfer between domains, and the contribution of Transfer Learning techniques (feature-representation-transfer based on Denoising Auto Encoders) for reducing the gap in the feature space distributions. Our results demonstrate an excellent cross-domain generalisation performance with and without feature representation transfer in both directions. In the case of music, cross-domain approaches outperformed intra-domain models for Valence estimation, whereas for Speech intra-domain models achieve the best performance. This is the first demonstration of shared acoustic codes for emotional expression in music and speech in the time-continuous domain.

摘要

音乐和言语在声学领域的情感交流方面表现出惊人的相似性,即特定情感的交流至少在一定程度上是通过共享的声学模式来实现的。从情感科学的角度来看,确定这两个领域之间的重叠程度对于理解这种现象背后的共享机制至关重要。从机器学习的角度来看,音乐和言语中情感表达的声学代码之间的重叠为扩大用于开发音乐和言语情感识别系统的数据量开辟了新的可能性。在本文中,我们研究了音乐和言语中情感(唤醒和效价)的时间连续预测,以及这两个领域之间的迁移学习。我们建立了一个比较框架,包括域内(即,在相同模态上训练和测试的模型,无论是音乐还是言语)和跨域实验(即,在一种模态上训练并在另一种模态上测试的模型)。在跨域背景下,我们评估了两种策略——域之间的直接迁移,以及迁移学习技术(基于去噪自动编码器的特征表示迁移)对缩小特征空间分布差距的贡献。我们的结果表明,无论是否进行双向特征表示迁移,都具有出色的跨域泛化性能。在音乐方面,跨域方法在效价估计方面优于域内模型,而对于言语,域内模型表现最佳。这是在时间连续域中音乐和言语情感表达共享声学代码的首次证明。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/adea/5489171/0c0b40546a0f/pone.0179289.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/adea/5489171/f53b6aa6b71c/pone.0179289.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/adea/5489171/2f7b33fbd6f6/pone.0179289.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/adea/5489171/48f4b40d1fe6/pone.0179289.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/adea/5489171/5c16c93ec598/pone.0179289.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/adea/5489171/dbb0c73952c3/pone.0179289.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/adea/5489171/0c0b40546a0f/pone.0179289.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/adea/5489171/f53b6aa6b71c/pone.0179289.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/adea/5489171/2f7b33fbd6f6/pone.0179289.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/adea/5489171/48f4b40d1fe6/pone.0179289.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/adea/5489171/5c16c93ec598/pone.0179289.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/adea/5489171/dbb0c73952c3/pone.0179289.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/adea/5489171/0c0b40546a0f/pone.0179289.g006.jpg

相似文献

1
Shared acoustic codes underlie emotional communication in music and speech-Evidence from deep transfer learning.共享声学编码是音乐和言语中情感交流的基础——来自深度迁移学习的证据。
PLoS One. 2017 Jun 28;12(6):e0179289. doi: 10.1371/journal.pone.0179289. eCollection 2017.
2
Psychoacoustic cues to emotion in speech prosody and music.言语韵律和音乐中的情感的心理声学线索。
Cogn Emot. 2013;27(4):658-84. doi: 10.1080/02699931.2012.732559. Epub 2012 Oct 12.
3
The time course of emotion recognition in speech and music.言语和音乐中的情绪识别时间进程。
J Acoust Soc Am. 2019 May;145(5):3058. doi: 10.1121/1.5108601.
4
On the Acoustics of Emotion in Audio: What Speech, Music, and Sound have in Common.音频情感声学:言语、音乐和声音的共同之处
Front Psychol. 2013 May 27;4:292. doi: 10.3389/fpsyg.2013.00292. eCollection 2013.
5
Temporal modulation in speech, music, and animal vocal communication: evidence of conserved function.言语、音乐和动物声音通讯中的时间调制:功能保守性的证据。
Ann N Y Acad Sci. 2019 Oct;1453(1):99-113. doi: 10.1111/nyas.14228. Epub 2019 Sep 4.
6
Affective Latent Representation of Acoustic and Lexical Features for Emotion Recognition.用于情感识别的声学和词汇特征的情感潜在表示。
Sensors (Basel). 2020 May 4;20(9):2614. doi: 10.3390/s20092614.
7
Beyond intensity: Spectral features effectively predict music-induced subjective arousal.超越强度:光谱特征有效预测音乐诱发的主观唤醒。
Q J Exp Psychol (Hove). 2014;67(7):1428-46. doi: 10.1080/17470218.2013.863954. Epub 2013 Dec 16.
8
Do Individual Differences Influence Moment-by-Moment Reports of Emotion Perceived in Music and Speech Prosody?个体差异是否会影响对音乐和言语韵律中所感知到的情绪的即时报告?
Front Behav Neurosci. 2018 Aug 27;12:184. doi: 10.3389/fnbeh.2018.00184. eCollection 2018.
9
Speaking to the trained ear: musical expertise enhances the recognition of emotions in speech prosody.对有训练的耳朵说话:音乐专业知识增强了对语音韵律中情感的识别。
Emotion. 2011 Oct;11(5):1021-31. doi: 10.1037/a0024521.
10
The minor third communicates sadness in speech, mirroring its use in music.在言语中,小三度音程传达出悲伤的情绪,这与它在音乐中的运用相呼应。
Emotion. 2010 Jun;10(3):335-48. doi: 10.1037/a0017928.

引用本文的文献

1
Improved emotion differentiation under reduced acoustic variability of speech in autism.自闭症患者的言语声音变异性降低,情绪辨别能力提高。
BMC Med. 2024 Mar 14;22(1):121. doi: 10.1186/s12916-024-03341-y.
2
Exploring the application of machine learning to expert evaluation of research impact.探索机器学习在专家评估研究影响力中的应用。
PLoS One. 2023 Aug 3;18(8):e0288469. doi: 10.1371/journal.pone.0288469. eCollection 2023.
3
Segmentation-Assisted Fully Convolutional Neural Network Enhances Deep Learning Performance to Identify Proliferative Diabetic Retinopathy.

本文引用的文献

1
Representation learning: a review and new perspectives.表示学习:综述与新视角。
IEEE Trans Pattern Anal Mach Intell. 2013 Aug;35(8):1798-828. doi: 10.1109/TPAMI.2013.50.
2
Sad music induces pleasant emotion.悲伤的音乐能诱发愉悦的情绪。
Front Psychol. 2013 Jun 13;4:311. doi: 10.3389/fpsyg.2013.00311. eCollection 2013.
3
From everyday emotions to aesthetic emotions: towards a unified theory of musical emotions.从日常情感到审美情感:走向音乐情感的统一理论。
分割辅助全卷积神经网络提高深度学习性能以识别增殖性糖尿病视网膜病变。
J Clin Med. 2023 Jan 3;12(1):385. doi: 10.3390/jcm12010385.
4
Correction: Shared acoustic codes underlie emotional communication in music and speech-Evidence from deep transfer learning.更正:共享声学编码是音乐和言语中情感交流的基础——来自深度迁移学习的证据。
PLoS One. 2018 Jan 19;13(1):e0191754. doi: 10.1371/journal.pone.0191754. eCollection 2018.
Phys Life Rev. 2013 Sep;10(3):235-66. doi: 10.1016/j.plrev.2013.05.008. Epub 2013 May 29.
4
On the Acoustics of Emotion in Audio: What Speech, Music, and Sound have in Common.音频情感声学:言语、音乐和声音的共同之处
Front Psychol. 2013 May 27;4:292. doi: 10.3389/fpsyg.2013.00292. eCollection 2013.
5
Psychoacoustic cues to emotion in speech prosody and music.言语韵律和音乐中的情感的心理声学线索。
Cogn Emot. 2013;27(4):658-84. doi: 10.1080/02699931.2012.732559. Epub 2012 Oct 12.
6
Musical emotions: predicting second-by-second subjective feelings of emotion from low-level psychoacoustic features and physiological measurements.音乐情感:从低水平的心理声学特征和生理测量数据预测音乐的逐秒主观情感感受。
Emotion. 2011 Aug;11(4):921-37. doi: 10.1037/a0024700.
7
Universal recognition of three basic emotions in music.音乐中三种基本情感的普遍认知。
Curr Biol. 2009 Apr 14;19(7):573-6. doi: 10.1016/j.cub.2009.02.058. Epub 2009 Mar 19.
8
Emotional responses to music: the need to consider underlying mechanisms.对音乐的情感反应:考虑潜在机制的必要性。
Behav Brain Sci. 2008 Oct;31(5):559-75; discussion 575-621. doi: 10.1017/S0140525X08005293.
9
Emotions evoked by the sound of music: characterization, classification, and measurement.音乐之声唤起的情感:特征描述、分类及测量
Emotion. 2008 Aug;8(4):494-521. doi: 10.1037/1528-3542.8.4.494.
10
Emotions over time: synchronicity and development of subjective, physiological, and facial affective reactions to music.情感随时间的变化:对音乐的主观、生理和面部情感反应的同步性与发展
Emotion. 2007 Nov;7(4):774-88. doi: 10.1037/1528-3542.7.4.774.