• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

低维习得特征空间定量个体和群体在声音曲目上的差异。

Low-dimensional learned feature spaces quantify individual and group differences in vocal repertoires.

机构信息

Department of Computer Science, Duke University, Durham, United States.

Center for Cognitive Neurobiology, Duke University, Durham, United States.

出版信息

Elife. 2021 May 14;10:e67855. doi: 10.7554/eLife.67855.

DOI:10.7554/eLife.67855
PMID:33988503
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8213406/
Abstract

Increases in the scale and complexity of behavioral data pose an increasing challenge for data analysis. A common strategy involves replacing entire behaviors with small numbers of handpicked, domain-specific features, but this approach suffers from several crucial limitations. For example, handpicked features may miss important dimensions of variability, and correlations among them complicate statistical testing. Here, by contrast, we apply the variational autoencoder (VAE), an unsupervised learning method, to learn features directly from data and quantify the vocal behavior of two model species: the laboratory mouse and the zebra finch. The VAE converges on a parsimonious representation that outperforms handpicked features on a variety of common analysis tasks, enables the measurement of moment-by-moment vocal variability on the timescale of tens of milliseconds in the zebra finch, provides strong evidence that mouse ultrasonic vocalizations do not cluster as is commonly believed, and captures the similarity of tutor and pupil birdsong with qualitatively higher fidelity than previous approaches. In all, we demonstrate the utility of modern unsupervised learning approaches to the quantification of complex and high-dimensional vocal behavior.

摘要

行为数据的规模和复杂性不断增加,给数据分析带来了越来越大的挑战。一种常见的策略是用少量手工挑选的、特定于领域的特征来替代整个行为,但这种方法存在几个关键的局限性。例如,手工挑选的特征可能会错过重要的可变性维度,并且它们之间的相关性会使统计检验复杂化。相比之下,我们在这里应用了变分自动编码器 (VAE),这是一种无监督学习方法,从数据中直接学习特征,并量化了两种模式生物的发声行为:实验鼠和斑胸草雀。VAE 收敛于一种简洁的表示,在各种常见的分析任务中表现优于手工挑选的特征,能够在斑马雀的几十毫秒时间尺度上测量每一刻的发声可变性,提供了有力的证据表明,老鼠的超声波叫声并不像通常认为的那样聚类,并且比以前的方法更准确地捕捉到了导师和学生鸟类歌声的相似性。总之,我们展示了现代无监督学习方法在量化复杂和高维发声行为方面的实用性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/27d6c55ea752/elife-67855-fig7-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/9ac390130a2e/elife-67855-fig1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/0de367099376/elife-67855-fig1-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/cdca383de308/elife-67855-fig2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/327d0b2eb360/elife-67855-fig2-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/1e6a28dda578/elife-67855-fig2-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/e82fa6769d1d/elife-67855-fig2-figsupp3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/2abe9d06e941/elife-67855-fig2-figsupp4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/0889afcb0283/elife-67855-fig2-figsupp5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/f329adaaaa8c/elife-67855-fig2-figsupp6.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/f7d91af3a745/elife-67855-fig3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/a9d2b99611a7/elife-67855-fig3-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/42307fc3864a/elife-67855-fig3-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/55ab71954bc5/elife-67855-fig3-figsupp3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/f3765b47e191/elife-67855-fig4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/1799744f54b2/elife-67855-fig4-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/748982e5cac5/elife-67855-fig4-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/17ea65bc1e66/elife-67855-fig4-figsupp3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/d45aec94434d/elife-67855-fig5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/b8b1a3d1e61b/elife-67855-fig5-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/3dd713d085c8/elife-67855-fig5-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/082da13a2ef7/elife-67855-fig5-figsupp3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/6b01145f2148/elife-67855-fig5-figsupp4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/f86fc872101d/elife-67855-fig5-figsupp5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/f16c1d9177b1/elife-67855-fig6.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/a0d277cb2292/elife-67855-fig6-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/5b6ec8e04c96/elife-67855-fig6-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/77fb25be5edb/elife-67855-fig7.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/3d4eed5ee82f/elife-67855-fig7-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/27d6c55ea752/elife-67855-fig7-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/9ac390130a2e/elife-67855-fig1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/0de367099376/elife-67855-fig1-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/cdca383de308/elife-67855-fig2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/327d0b2eb360/elife-67855-fig2-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/1e6a28dda578/elife-67855-fig2-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/e82fa6769d1d/elife-67855-fig2-figsupp3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/2abe9d06e941/elife-67855-fig2-figsupp4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/0889afcb0283/elife-67855-fig2-figsupp5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/f329adaaaa8c/elife-67855-fig2-figsupp6.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/f7d91af3a745/elife-67855-fig3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/a9d2b99611a7/elife-67855-fig3-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/42307fc3864a/elife-67855-fig3-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/55ab71954bc5/elife-67855-fig3-figsupp3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/f3765b47e191/elife-67855-fig4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/1799744f54b2/elife-67855-fig4-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/748982e5cac5/elife-67855-fig4-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/17ea65bc1e66/elife-67855-fig4-figsupp3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/d45aec94434d/elife-67855-fig5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/b8b1a3d1e61b/elife-67855-fig5-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/3dd713d085c8/elife-67855-fig5-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/082da13a2ef7/elife-67855-fig5-figsupp3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/6b01145f2148/elife-67855-fig5-figsupp4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/f86fc872101d/elife-67855-fig5-figsupp5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/f16c1d9177b1/elife-67855-fig6.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/a0d277cb2292/elife-67855-fig6-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/5b6ec8e04c96/elife-67855-fig6-figsupp2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/77fb25be5edb/elife-67855-fig7.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/3d4eed5ee82f/elife-67855-fig7-figsupp1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/634b/8213406/27d6c55ea752/elife-67855-fig7-figsupp2.jpg

相似文献

1
Low-dimensional learned feature spaces quantify individual and group differences in vocal repertoires.低维习得特征空间定量个体和群体在声音曲目上的差异。
Elife. 2021 May 14;10:e67855. doi: 10.7554/eLife.67855.
2
Songbirds can learn flexible contextual control over syllable sequencing.鸣禽可以学习灵活的上下文控制来调整音节序列。
Elife. 2021 Jun 1;10:e61610. doi: 10.7554/eLife.61610.
3
Finding, visualizing, and quantifying latent structure across diverse animal vocal repertoires.发现、可视化和量化不同动物声谱中的潜在结构。
PLoS Comput Biol. 2020 Oct 15;16(10):e1008228. doi: 10.1371/journal.pcbi.1008228. eCollection 2020 Oct.
4
An automated procedure for evaluating song imitation.一种评估歌曲模仿的自动化程序。
PLoS One. 2014 May 8;9(5):e96484. doi: 10.1371/journal.pone.0096484. eCollection 2014.
5
Quantitative assessment of vocal development in the zebra finch using self-organizing neural networks.使用自组织神经网络对斑胸草雀的发声发育进行定量评估。
J Acoust Soc Am. 2001 Nov;110(5 Pt 1):2593-603. doi: 10.1121/1.1412446.
6
Learning to pause: Fidelity of and biases in the developmental acquisition of gaps in the communicative signals of a songbird.学习停顿:鸣禽交际信号间隙发展习得中的保真度和偏差。
Dev Sci. 2023 Sep;26(5):e13382. doi: 10.1111/desc.13382. Epub 2023 Mar 2.
7
miR-9 regulates basal ganglia-dependent developmental vocal learning and adult vocal performance in songbirds.微小RNA-9调节鸣禽中基底神经节依赖的发育性发声学习及成年后的发声表现。
Elife. 2018 Jan 18;7:e29087. doi: 10.7554/eLife.29087.
8
Dynamical origin of spectrally rich vocalizations in birdsong.鸟鸣中频谱丰富发声的动态起源。
Phys Rev E Stat Nonlin Soft Matter Phys. 2008 Jul;78(1 Pt 1):011905. doi: 10.1103/PhysRevE.78.011905. Epub 2008 Jul 11.
9
Ability to modulate birdsong across social contexts develops without imitative social learning.无需模仿性的社会学习,鸟类也能在不同的社会环境中调节鸣叫声。
Biol Lett. 2018 Mar;14(3). doi: 10.1098/rsbl.2017.0777.
10
Introductory gestures before songbird vocal displays are shaped by learning and biological predispositions.鸣禽在进行鸣叫表演之前的介绍性动作是通过学习和生物倾向形成的。
Proc Biol Sci. 2021 Jan 27;288(1943):20202796. doi: 10.1098/rspb.2020.2796. Epub 2021 Jan 20.

引用本文的文献

1
Correctness is its own reward: bootstrapping error signals in self-guided reinforcement learning.正确性本身就是一种回报:在自我引导的强化学习中引导误差信号。
bioRxiv. 2025 Aug 19:2025.07.18.665446. doi: 10.1101/2025.07.18.665446.
2
Hidden in plain sound: the scientific potential of house mouse squeaks.隐藏于平常声音之中:家鼠叫声的科学潜力
Biol Lett. 2025 Aug;21(8):20250333. doi: 10.1098/rsbl.2025.0333. Epub 2025 Aug 27.
3
Model Predictive Control on the Neural Manifold.神经流形上的模型预测控制

本文引用的文献

1
Finding, visualizing, and quantifying latent structure across diverse animal vocal repertoires.发现、可视化和量化不同动物声谱中的潜在结构。
PLoS Comput Biol. 2020 Oct 15;16(10):e1008228. doi: 10.1371/journal.pcbi.1008228. eCollection 2020 Oct.
2
Temporal structure of mouse courtship vocalizations facilitates syllable labeling.老鼠求偶叫声的时间结构有助于音节标记。
Commun Biol. 2020 Jun 26;3(1):333. doi: 10.1038/s42003-020-1053-7.
3
Classifying sex and strain from mouse ultrasonic vocalizations using deep learning.使用深度学习对小鼠超声发声进行性别和品系分类。
ArXiv. 2025 Aug 11:arXiv:2406.14801v2.
4
Computational Urban Ecology of New York City Rats.纽约市老鼠的计算城市生态学
bioRxiv. 2025 Jul 24:2025.07.21.665423. doi: 10.1101/2025.07.21.665423.
5
Infant rat ultrasonic vocalizations in the neurodevelopmental model of schizophrenia.精神分裂症神经发育模型中的幼鼠超声波发声
Sci Rep. 2025 Jul 28;15(1):27472. doi: 10.1038/s41598-025-08412-5.
6
Convergent vocal representations in parrot and human forebrain motor networks.鹦鹉和人类前脑运动网络中的趋同语音表征。
Nature. 2025 Apr;640(8058):427-434. doi: 10.1038/s41586-025-08695-8. Epub 2025 Mar 19.
7
Dual neuromodulatory dynamics underlie birdsong learning.双重神经调节动力学是鸟鸣学习的基础。
Nature. 2025 May;641(8063):690-698. doi: 10.1038/s41586-025-08694-9. Epub 2025 Mar 12.
8
Sensory input, sex and function shape hypothalamic cell type development.感觉输入、性别和功能塑造下丘脑细胞类型的发育。
Nature. 2025 Mar 5. doi: 10.1038/s41586-025-08603-0.
9
The utility of animal models to inform the next generation of human space exploration.动物模型对指导下一代人类太空探索的效用。
NPJ Microgravity. 2025 Feb 22;11(1):7. doi: 10.1038/s41526-025-00460-5.
10
Detection of Individual Differences Encoded in Sequential Variations of Elements in Zebra Finch Songs.斑胸草雀歌声中元素序列变化所编码的个体差异的检测。
J Neurosci. 2025 Apr 2;45(14):e1071242025. doi: 10.1523/JNEUROSCI.1071-24.2025.
PLoS Comput Biol. 2020 Jun 22;16(6):e1007918. doi: 10.1371/journal.pcbi.1007918. eCollection 2020 Jun.
4
Nearest neighbours reveal fast and slow components of motor learning.最近邻揭示了运动学习的快和慢成分。
Nature. 2020 Jan;577(7791):526-530. doi: 10.1038/s41586-019-1892-x. Epub 2020 Jan 8.
5
Discovering Precise Temporal Patterns in Large-Scale Neural Recordings through Robust and Interpretable Time Warping.通过稳健且可解释的时间 warp 发现大规模神经记录中的精确时间模式。
Neuron. 2020 Jan 22;105(2):246-259.e8. doi: 10.1016/j.neuron.2019.10.020. Epub 2019 Nov 27.
6
Parallels in the sequential organization of birdsong and human speech.鸟鸣和人类言语在顺序组织上的相似性。
Nat Commun. 2019 Aug 12;10(1):3636. doi: 10.1038/s41467-019-11605-y.
7
DeepSqueak: a deep learning-based system for detection and analysis of ultrasonic vocalizations.DeepSqueak:一种基于深度学习的超声发声检测与分析系统。
Neuropsychopharmacology. 2019 Apr;44(5):859-868. doi: 10.1038/s41386-018-0303-6. Epub 2019 Jan 4.
8
An automated approach to the quantitation of vocalizations and vocal learning in the songbird.一种自动分析鸟类发声和学习的方法。
PLoS Comput Biol. 2018 Aug 31;14(8):e1006437. doi: 10.1371/journal.pcbi.1006437. eCollection 2018 Aug.
9
MUPET-Mouse Ultrasonic Profile ExTraction: A Signal Processing Tool for Rapid and Unsupervised Analysis of Ultrasonic Vocalizations.MUPET-小鼠超声轮廓提取:一种用于快速且无监督分析超声发声的信号处理工具。
Neuron. 2017 May 3;94(3):465-485.e5. doi: 10.1016/j.neuron.2017.04.005.
10
Efficient "Shotgun" Inference of Neural Connectivity from Highly Sub-sampled Activity Data.从高度下采样的活动数据中高效“散弹枪式”推断神经连接性
PLoS Comput Biol. 2015 Oct 14;11(10):e1004464. doi: 10.1371/journal.pcbi.1004464. eCollection 2015 Oct.