• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通用头部相关传递函数在虚拟现实中可能就足够好了。通过跨模态可塑性改善声源定位。

Generic HRTFs May be Good Enough in Virtual Reality. Improving Source Localization through Cross-Modal Plasticity.

作者信息

Berger Christopher C, Gonzalez-Franco Mar, Tajadura-Jiménez Ana, Florencio Dinei, Zhang Zhengyou

机构信息

Microsoft Research, Redmond, WA, United States.

Division of Biology and Biological Engineering, California Institute of Technology, Pasadena, CA, United States.

出版信息

Front Neurosci. 2018 Feb 2;12:21. doi: 10.3389/fnins.2018.00021. eCollection 2018.

DOI:10.3389/fnins.2018.00021
PMID:29456486
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC5801410/
Abstract

Auditory spatial localization in humans is performed using a combination of interaural time differences, interaural level differences, as well as spectral cues provided by the geometry of the ear. To render spatialized sounds within a virtual reality (VR) headset, either individualized or generic Head Related Transfer Functions (HRTFs) are usually employed. The former require arduous calibrations, but enable accurate auditory source localization, which may lead to a heightened sense of presence within VR. The latter obviate the need for individualized calibrations, but result in less accurate auditory source localization. Previous research on auditory source localization in the real world suggests that our representation of acoustic space is highly plastic. In light of these findings, we investigated whether auditory source localization could be improved for users of generic HRTFs via cross-modal learning. The results show that pairing a dynamic auditory stimulus, with a spatio-temporally aligned visual counterpart, enabled users of generic HRTFs to improve subsequent auditory source localization. Exposure to the auditory stimulus alone or to asynchronous audiovisual stimuli did not improve auditory source localization. These findings have important implications for human perception as well as the development of VR systems as they indicate that generic HRTFs may be enough to enable good auditory source localization in VR.

摘要

人类的听觉空间定位是通过双耳时间差、双耳声级差以及耳朵几何形状提供的频谱线索的组合来完成的。为了在虚拟现实(VR)头戴式设备中呈现空间化声音,通常会采用个性化或通用的头部相关传递函数(HRTF)。前者需要进行艰巨的校准,但能够实现精确的声源定位,这可能会增强VR中的临场感。后者无需进行个性化校准,但会导致声源定位不够准确。先前关于现实世界中声源定位的研究表明,我们对声学空间的表征具有高度可塑性。鉴于这些发现,我们研究了是否可以通过跨模态学习来改善通用HRTF用户的声源定位。结果表明,将动态听觉刺激与时空对齐的视觉对应物配对,能够使通用HRTF用户改善后续的声源定位。仅接触听觉刺激或异步视听刺激并不能改善声源定位。这些发现对人类感知以及VR系统的开发具有重要意义,因为它们表明通用HRTF可能足以在VR中实现良好的声源定位。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9849/5801410/10b1fef191ec/fnins-12-00021-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9849/5801410/b7a2f35c7103/fnins-12-00021-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9849/5801410/10b1fef191ec/fnins-12-00021-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9849/5801410/b7a2f35c7103/fnins-12-00021-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9849/5801410/10b1fef191ec/fnins-12-00021-g0002.jpg

相似文献

1
Generic HRTFs May be Good Enough in Virtual Reality. Improving Source Localization through Cross-Modal Plasticity.通用头部相关传递函数在虚拟现实中可能就足够好了。通过跨模态可塑性改善声源定位。
Front Neurosci. 2018 Feb 2;12:21. doi: 10.3389/fnins.2018.00021. eCollection 2018.
2
Usability of Individualized Head-Related Transfer Functions in Virtual Reality: Empirical Study With Perceptual Attributes in Sagittal Plane Sound Localization.虚拟现实中个性化头部相关传递函数的可用性:矢状面声音定位中感知属性的实证研究。
JMIR Serious Games. 2020 Sep 8;8(3):e17576. doi: 10.2196/17576.
3
Cross-modal correspondence enhances elevation localization in visual-to-auditory sensory substitution.跨模态对应增强了视觉到听觉感官替代中的高度定位。
Front Psychol. 2023 Jan 26;14:1079998. doi: 10.3389/fpsyg.2023.1079998. eCollection 2023.
4
Improvements of sound localization abilities by the facial ruff of the barn owl (Tyto alba) as demonstrated by virtual ruff removal.通过虚拟去除面羽,展示了仓鸮(Tyto alba)的面部羽领对面部声音定位能力的改善。
PLoS One. 2009 Nov 5;4(11):e7721. doi: 10.1371/journal.pone.0007721.
5
The effect of head-related transfer function measurement methodology on localization performance in spatial audio interfaces.头部相关传递函数测量方法对空间音频接口中定位性能的影响。
Hum Factors. 2008 Apr;50(2):256-63. doi: 10.1518/001872008X250700.
6
Head-related transfer functions of rabbits within the front horizontal plane.兔在额状面内的头部相关传递函数。
Hear Res. 2024 Jan;441:108924. doi: 10.1016/j.heares.2023.108924. Epub 2023 Dec 2.
7
The role of individualized headphone calibration for the generation of high fidelity virtual auditory space.个性化耳机校准在生成高保真虚拟听觉空间中的作用。
J Acoust Soc Am. 1996 Dec;100(6):3785-93. doi: 10.1121/1.417337.
8
Sound source localization with varying amount of visual information in virtual reality.虚拟现实中具有不同数量视觉信息的声源定位。
PLoS One. 2019 Mar 29;14(3):e0214603. doi: 10.1371/journal.pone.0214603. eCollection 2019.
9
The effect of different cochlear implant microphones on acoustic hearing individuals' binaural benefits for speech perception in noise.不同人工耳蜗麦克风对听力正常者噪声环境下言语感知的双侧获益的影响。
Ear Hear. 2011 Jul-Aug;32(4):468-84. doi: 10.1097/AUD.0b013e31820dd3f0.
10
Comparative Analysis of HRTFs Measurement Using In-Ear Microphones.使用入耳式麦克风的 HRTFs 测量比较分析。
Sensors (Basel). 2023 Jun 29;23(13):6016. doi: 10.3390/s23136016.

引用本文的文献

1
Age-Related Differences in Neural Correlates of Auditory Spatial Change Detection in Real and Virtual Environments.真实和虚拟环境中听觉空间变化检测的神经关联的年龄相关差异
Eur J Neurosci. 2025 May;61(10):e70141. doi: 10.1111/ejn.70141.
2
Multisensory stimuli facilitate low-level perceptual learning on a difficult global motion task in virtual reality.多感官刺激促进虚拟现实中一项困难的全局运动任务的低水平感知学习。
PLoS One. 2025 Mar 4;20(3):e0319007. doi: 10.1371/journal.pone.0319007. eCollection 2025.
3
Happy new ears: Rapid adaptation to novel spectral cues in vertical sound localization.

本文引用的文献

1
Mental Imagery Induces Cross-Modal Sensory Plasticity and Changes Future Auditory Perception.心理意象诱导跨模态感觉可塑性并改变未来的听觉感知。
Psychol Sci. 2018 Jun;29(6):926-935. doi: 10.1177/0956797617748959. Epub 2018 Apr 10.
2
Model of Illusions and Virtual Reality.幻觉与虚拟现实模型
Front Psychol. 2017 Jun 30;8:1125. doi: 10.3389/fpsyg.2017.01125. eCollection 2017.
3
Concurrent talking in immersive virtual reality: on the dominance of visual speech cues.沉浸式虚拟现实中的同时对话:关于视觉言语线索的主导地位。
新年快乐:垂直声音定位中对新频谱线索的快速适应
iScience. 2024 Nov 4;27(12):111308. doi: 10.1016/j.isci.2024.111308. eCollection 2024 Dec 20.
4
Auditory localization: a comprehensive practical review.听觉定位:全面的实践综述。
Front Psychol. 2024 Jul 10;15:1408073. doi: 10.3389/fpsyg.2024.1408073. eCollection 2024.
5
Sound localization in web-based 3D environments.基于网络的 3D 环境中的声音定位。
Sci Rep. 2022 Jul 15;12(1):12107. doi: 10.1038/s41598-022-15931-y.
6
Virtual reality training improves dynamic balance in children with cerebral palsy.虚拟现实训练可改善脑瘫患儿的动态平衡能力。
Int J Dev Disabil. 2019 Oct 26;67(6):429-434. doi: 10.1080/20473869.2019.1679471. eCollection 2021.
7
Short-Term Audiovisual Spatial Training Enhances Electrophysiological Correlates of Auditory Selective Spatial Attention.短期视听空间训练增强听觉选择性空间注意的电生理相关性。
Front Neurosci. 2021 Jul 1;15:645702. doi: 10.3389/fnins.2021.645702. eCollection 2021.
8
Investigation of an MAA Test With Virtual Sound Synthesis.虚拟声音合成的MAA测试研究。
Front Psychol. 2021 Jun 2;12:656052. doi: 10.3389/fpsyg.2021.656052. eCollection 2021.
9
Visual Influences on Auditory Behavioral, Neural, and Perceptual Processes: A Review.视觉对听觉行为、神经和知觉过程的影响:综述。
J Assoc Res Otolaryngol. 2021 Jul;22(4):365-386. doi: 10.1007/s10162-021-00789-0. Epub 2021 May 20.
10
Rethinking GPS navigation: creating cognitive maps through auditory clues.重新思考 GPS 导航:通过听觉线索创建认知地图。
Sci Rep. 2021 Apr 8;11(1):7764. doi: 10.1038/s41598-021-87148-4.
Sci Rep. 2017 Jun 19;7(1):3817. doi: 10.1038/s41598-017-04201-x.
4
The Plausibility of a String Quartet Performance in Virtual Reality.虚拟现实中弦乐四重奏表演的可能性。
IEEE Trans Vis Comput Graph. 2017 Apr;23(4):1352-1359. doi: 10.1109/TVCG.2017.2657138. Epub 2017 Jan 27.
5
Auditory Motion Elicits a Visual Motion Aftereffect.听觉运动引发视觉运动后效。
Front Neurosci. 2016 Dec 5;10:559. doi: 10.3389/fnins.2016.00559. eCollection 2016.
6
Violating body movement semantics: Neural signatures of self-generated and external-generated errors.违反身体运动语义:自我产生和外部产生错误的神经特征。
Neuroimage. 2016 Jan 1;124(Pt A):147-156. doi: 10.1016/j.neuroimage.2015.08.022. Epub 2015 Aug 15.
7
Multisensory perception as an associative learning process.多感官知觉作为一种联想学习过程。
Front Psychol. 2014 Sep 26;5:1095. doi: 10.3389/fpsyg.2014.01095. eCollection 2014.
8
The plastic ear and perceptual relearning in auditory spatial perception.塑料耳朵与听觉空间感知中的知觉再学习。
Front Neurosci. 2014 Aug 6;8:237. doi: 10.3389/fnins.2014.00237. eCollection 2014.
9
Acoustic and non-acoustic factors in modeling listener-specific performance of sagittal-plane sound localization.在矢状面声音定位中建模听众特定性能的声学和非声学因素。
Front Psychol. 2014 Apr 23;5:319. doi: 10.3389/fpsyg.2014.00319. eCollection 2014.
10
Intersensory binding across space and time: a tutorial review.跨时空的多感官整合:教程综述
Atten Percept Psychophys. 2013 Jul;75(5):790-811. doi: 10.3758/s13414-013-0475-4.