• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

听觉分类的最佳特征。

Optimal features for auditory categorization.

机构信息

Department of Bioengineering, University of Pittsburgh, Pittsburgh, 15213, PA, USA.

Department of Neurobiology, University of Pittsburgh, Pittsburgh, 15213, PA, USA.

出版信息

Nat Commun. 2019 Mar 21;10(1):1302. doi: 10.1038/s41467-019-09115-y.

DOI:10.1038/s41467-019-09115-y
PMID:30899018
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC6428858/
Abstract

Humans and vocal animals use vocalizations to communicate with members of their species. A necessary function of auditory perception is to generalize across the high variability inherent in vocalization production and classify them into behaviorally distinct categories ('words' or 'call types'). Here, we demonstrate that detecting mid-level features in calls achieves production-invariant classification. Starting from randomly chosen marmoset call features, we use a greedy search algorithm to determine the most informative and least redundant features necessary for call classification. High classification performance is achieved using only 10-20 features per call type. Predictions of tuning properties of putative feature-selective neurons accurately match some observed auditory cortical responses. This feature-based approach also succeeds for call categorization in other species, and for other complex classification tasks such as caller identification. Our results suggest that high-level neural representations of sounds are based on task-dependent features optimized for specific computational goals.

摘要

人类和发声动物利用发声来与同种成员进行交流。听觉感知的一个必要功能是对发声产生中的高度可变性进行概括,并将其分类为具有不同行为特征的类别(“单词”或“叫声类型”)。在这里,我们证明了在叫声中检测中间层特征可实现与产生无关的分类。从随机选择的狨猴叫声特征开始,我们使用贪婪搜索算法来确定对于叫声分类最有用且最不冗余的特征。每个叫声类型仅使用 10-20 个特征即可实现高分类性能。对候选特征选择神经元调谐特性的预测与一些观察到的听觉皮层反应非常吻合。这种基于特征的方法也适用于其他物种的叫声分类,以及其他复杂的分类任务,如呼叫者识别。我们的研究结果表明,声音的高级神经表示是基于针对特定计算目标优化的任务相关特征。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/c24e3b63ee94/41467_2019_9115_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/b8787c3aa5af/41467_2019_9115_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/9ecf9fbd00ab/41467_2019_9115_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/b43df911e61a/41467_2019_9115_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/cd74da8ecbfd/41467_2019_9115_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/d5fde566be53/41467_2019_9115_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/eecd1fc5b648/41467_2019_9115_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/20a45dadd899/41467_2019_9115_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/cff45c14f708/41467_2019_9115_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/bb9782fe530c/41467_2019_9115_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/c24e3b63ee94/41467_2019_9115_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/b8787c3aa5af/41467_2019_9115_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/9ecf9fbd00ab/41467_2019_9115_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/b43df911e61a/41467_2019_9115_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/cd74da8ecbfd/41467_2019_9115_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/d5fde566be53/41467_2019_9115_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/eecd1fc5b648/41467_2019_9115_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/20a45dadd899/41467_2019_9115_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/cff45c14f708/41467_2019_9115_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/bb9782fe530c/41467_2019_9115_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/0afc/6428858/c24e3b63ee94/41467_2019_9115_Fig10_HTML.jpg

相似文献

1
Optimal features for auditory categorization.听觉分类的最佳特征。
Nat Commun. 2019 Mar 21;10(1):1302. doi: 10.1038/s41467-019-09115-y.
2
Virtual vocalization stimuli for investigating neural representations of species-specific vocalizations.用于研究物种特异性发声神经表征的虚拟发声刺激。
J Neurophysiol. 2006 Feb;95(2):1244-62. doi: 10.1152/jn.00818.2005. Epub 2005 Oct 5.
3
Plasticity of temporal pattern codes for vocalization stimuli in primary auditory cortex.初级听觉皮层中发声刺激的时间模式编码可塑性。
J Neurosci. 2006 May 3;26(18):4785-95. doi: 10.1523/JNEUROSCI.4330-05.2006.
4
Contributions of sensory tuning to auditory-vocal interactions in marmoset auditory cortex.感觉调谐对狨猴听觉皮层中听觉-发声相互作用的贡献。
Hear Res. 2017 May;348:98-111. doi: 10.1016/j.heares.2017.03.001. Epub 2017 Mar 9.
5
Auditory Selectivity for Spectral Contrast in Cortical Neurons and Behavior.皮层神经元和行为的光谱对比听觉选择性。
J Neurosci. 2020 Jan 29;40(5):1015-1027. doi: 10.1523/JNEUROSCI.1200-19.2019. Epub 2019 Dec 11.
6
Representation of a species-specific vocalization in the primary auditory cortex of the common marmoset: temporal and spectral characteristics.普通狨猴初级听觉皮层中一种物种特异性发声的表征:时间和频谱特征。
J Neurophysiol. 1995 Dec;74(6):2685-706. doi: 10.1152/jn.1995.74.6.2685.
7
Vocalization categorization behavior explained by a feature-based auditory categorization model.基于特征的听觉分类模型解释发声分类行为。
Elife. 2022 Oct 13;11:e78278. doi: 10.7554/eLife.78278.
8
Contextual effects of noise on vocalization encoding in primary auditory cortex.噪声对初级听觉皮层中发声编码的情境效应。
J Neurophysiol. 2017 Feb 1;117(2):713-727. doi: 10.1152/jn.00476.2016. Epub 2016 Nov 23.
9
Representation of spectral and temporal envelope of twitter vocalizations in common marmoset primary auditory cortex.普通狨猴初级听觉皮层中推特叫声的频谱和时间包络的表征。
J Neurophysiol. 2002 Apr;87(4):1723-37. doi: 10.1152/jn.00632.2001.
10
Neuronal selectivity to complex vocalization features emerges in the superficial layers of primary auditory cortex.对复杂发声特征的神经元选择性出现在初级听觉皮层的浅层。
PLoS Biol. 2021 Jun 16;19(6):e3001299. doi: 10.1371/journal.pbio.3001299. eCollection 2021 Jun.

引用本文的文献

1
Spatially clustered neurons in the bat midbrain encode vocalization categories.蝙蝠中脑中空间聚集的神经元对发声类别进行编码。
Nat Neurosci. 2025 May;28(5):1038-1047. doi: 10.1038/s41593-025-01932-3. Epub 2025 Apr 14.
2
Systematic changes in neural selectivity reflect the acquired salience of category-diagnostic dimensions.神经选择性的系统性变化反映了类别诊断维度所获得的显著性。
bioRxiv. 2024 Sep 23:2024.09.21.614258. doi: 10.1101/2024.09.21.614258.
3
Representation of conspecific vocalizations in amygdala of awake marmosets.清醒狨猴杏仁核中同种发声的表征

本文引用的文献

1
Sound identity is represented robustly in auditory cortex during perceptual constancy.声音身份在知觉恒常性期间在听觉皮层中得到了强有力的表现。
Nat Commun. 2018 Nov 14;9(1):4786. doi: 10.1038/s41467-018-07237-3.
2
A Task-Optimized Neural Network Replicates Human Auditory Behavior, Predicts Brain Responses, and Reveals a Cortical Processing Hierarchy.任务优化神经网络复制人类听觉行为,预测大脑反应,并揭示皮质处理层次结构。
Neuron. 2018 May 2;98(3):630-644.e16. doi: 10.1016/j.neuron.2018.03.044. Epub 2018 Apr 19.
3
Training Humans to Categorize Monkey Calls: Auditory Feature- and Category-Selective Neural Tuning Changes.
Natl Sci Rev. 2023 Jul 13;10(11):nwad194. doi: 10.1093/nsr/nwad194. eCollection 2023 Nov.
4
The neurobiology of vocal communication in marmosets.狨猴的发声通讯神经生物学。
Ann N Y Acad Sci. 2023 Oct;1528(1):13-28. doi: 10.1111/nyas.15057. Epub 2023 Aug 24.
5
Spatially clustered neurons encode vocalization categories in the bat midbrain.空间聚集的神经元在蝙蝠中脑对发声类别进行编码。
bioRxiv. 2023 Jun 14:2023.06.14.545029. doi: 10.1101/2023.06.14.545029.
6
Adaptive mechanisms facilitate robust performance in noise and in reverberation in an auditory categorization model.自适应机制使听觉分类模型在噪声和混响中具有强大的性能。
Commun Biol. 2023 May 2;6(1):456. doi: 10.1038/s42003-023-04816-z.
7
Quantitative models of auditory cortical processing.听觉皮层处理的定量模型。
Hear Res. 2023 Mar 1;429:108697. doi: 10.1016/j.heares.2023.108697. Epub 2023 Jan 14.
8
Relative pitch representations and invariance to timbre.相对音高表示法与音色不变性。
Cognition. 2023 Mar;232:105327. doi: 10.1016/j.cognition.2022.105327. Epub 2022 Dec 7.
9
Vocalization categorization behavior explained by a feature-based auditory categorization model.基于特征的听觉分类模型解释发声分类行为。
Elife. 2022 Oct 13;11:e78278. doi: 10.7554/eLife.78278.
10
Principal component decomposition of acoustic and neural representations of time-varying pitch reveals adaptive efficient coding of speech covariation patterns.时变音高的声学分和神经表示的主成分分解揭示了言语协变模式的自适应有效编码。
Brain Lang. 2022 Jul;230:105122. doi: 10.1016/j.bandl.2022.105122. Epub 2022 Apr 20.
训练人类对猴子叫声进行分类:听觉特征和类别选择性神经调谐变化。
Neuron. 2018 Apr 18;98(2):405-416.e4. doi: 10.1016/j.neuron.2018.03.014.
4
Analyzing Distributional Learning of Phonemic Categories in Unsupervised Deep Neural Networks.分析无监督深度神经网络中音素类别的分布学习
Cogsci. 2016 Aug;2016:1757-1762.
5
Learning Midlevel Auditory Codes from Natural Sound Statistics.从自然声音统计中学习中级听觉编码。
Neural Comput. 2018 Mar;30(3):631-669. doi: 10.1162/neco_a_01048. Epub 2017 Dec 8.
6
Familiarity and Within-Person Facial Variability: The Importance of the Internal and External Features.熟悉度与个体面部变异性:内部和外部特征的重要性。
Perception. 2018 Jan;47(1):3-15. doi: 10.1177/0301006617725242. Epub 2017 Aug 13.
7
Dynamic Encoding of Acoustic Features in Neural Responses to Continuous Speech.对连续语音的神经反应中声学特征的动态编码
J Neurosci. 2017 Feb 22;37(8):2176-2185. doi: 10.1523/JNEUROSCI.2383-16.2017. Epub 2017 Jan 24.
8
Distributed acoustic cues for caller identity in macaque vocalization.猴叫声中用于识别来电者身份的分布式声学线索。
R Soc Open Sci. 2015 Dec 23;2(12):150432. doi: 10.1098/rsos.150432. eCollection 2015 Dec.
9
Atoms of recognition in human and computer vision.人类视觉与计算机视觉中的识别原子。
Proc Natl Acad Sci U S A. 2016 Mar 8;113(10):2744-9. doi: 10.1073/pnas.1513198113. Epub 2016 Feb 16.
10
A quantitative acoustic analysis of the vocal repertoire of the common marmoset (Callithrix jacchus).普通狨猴(Callithrix jacchus)发声库的定量声学分析。
J Acoust Soc Am. 2015 Nov;138(5):2906-28. doi: 10.1121/1.4934268.