• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

人类视觉系统和卷积神经网络都可以支持在经历极端位移后实现稳健的在线翻译容忍度。

The human visual system and CNNs can both support robust online translation tolerance following extreme displacements.

机构信息

School of Psychological Science, University of Bristol, Bristol, UK.

Department of Cognitive Science and Psychology, Sofia, New Bulgarian University, Bulgaria.

出版信息

J Vis. 2021 Feb 3;21(2):9. doi: 10.1167/jov.21.2.9.

DOI:10.1167/jov.21.2.9
PMID:33620380
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7910631/
Abstract

Visual translation tolerance refers to our capacity to recognize objects over a wide range of different retinal locations. Although translation is perhaps the simplest spatial transform that the visual system needs to cope with, the extent to which the human visual system can identify objects at previously unseen locations is unclear, with some studies reporting near complete invariance over 10 degrees and other reporting zero invariance at 4 degrees of visual angle. Similarly, there is confusion regarding the extent of translation tolerance in computational models of vision, as well as the degree of match between human and model performance. Here, we report a series of eye-tracking studies (total N = 70) demonstrating that novel objects trained at one retinal location can be recognized at high accuracy rates following translations up to 18 degrees. We also show that standard deep convolutional neural networks (DCNNs) support our findings when pretrained to classify another set of stimuli across a range of locations, or when a global average pooling (GAP) layer is added to produce larger receptive fields. Our findings provide a strong constraint for theories of human vision and help explain inconsistent findings previously reported with convolutional neural networks (CNNs).

摘要

视觉平移容忍度指的是我们在大范围不同视网膜位置识别物体的能力。尽管平移可能是视觉系统需要应对的最简单的空间变换,但人类视觉系统在以前未见过的位置识别物体的程度尚不清楚,一些研究报告称在 10 度以上几乎完全不变,而其他研究则报告在 4 度视角下完全不变。同样,在视觉计算模型中,平移容忍度的程度以及人类和模型性能之间的匹配程度也存在混淆。在这里,我们报告了一系列眼动追踪研究(总 N = 70),这些研究表明,在一个视网膜位置上训练的新物体可以在高达 18 度的平移后以高准确率识别。我们还表明,当标准的深度卷积神经网络(DCNN)经过预训练以在一系列位置上对另一组刺激进行分类时,或者当添加全局平均池化(GAP)层以产生更大的感受野时,该网络支持我们的发现。我们的研究结果为人类视觉理论提供了强有力的约束,并有助于解释以前使用卷积神经网络(CNN)报告的不一致发现。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3cc3/7910631/7e97a2917fef/jovi-21-2-9-f005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3cc3/7910631/0ebdc131dc86/jovi-21-2-9-f001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3cc3/7910631/549c63f5be08/jovi-21-2-9-f002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3cc3/7910631/6c9096e6ded3/jovi-21-2-9-f003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3cc3/7910631/8711ad54412f/jovi-21-2-9-f004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3cc3/7910631/7e97a2917fef/jovi-21-2-9-f005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3cc3/7910631/0ebdc131dc86/jovi-21-2-9-f001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3cc3/7910631/549c63f5be08/jovi-21-2-9-f002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3cc3/7910631/6c9096e6ded3/jovi-21-2-9-f003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3cc3/7910631/8711ad54412f/jovi-21-2-9-f004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3cc3/7910631/7e97a2917fef/jovi-21-2-9-f005.jpg

相似文献

1
The human visual system and CNNs can both support robust online translation tolerance following extreme displacements.人类视觉系统和卷积神经网络都可以支持在经历极端位移后实现稳健的在线翻译容忍度。
J Vis. 2021 Feb 3;21(2):9. doi: 10.1167/jov.21.2.9.
2
The visual system supports online translation invariance for object identification.视觉系统支持用于物体识别的在线平移不变性。
Psychon Bull Rev. 2016 Apr;23(2):432-8. doi: 10.3758/s13423-015-0916-2.
3
Crowding in humans is unlike that in convolutional neural networks.人群拥挤的情况与卷积神经网络不同。
Neural Netw. 2020 Jun;126:262-274. doi: 10.1016/j.neunet.2020.03.021. Epub 2020 Mar 27.
4
Deep convolutional networks do not classify based on global object shape.深度卷积网络不是基于全局物体形状进行分类的。
PLoS Comput Biol. 2018 Dec 7;14(12):e1006613. doi: 10.1371/journal.pcbi.1006613. eCollection 2018 Dec.
5
Improved object recognition using neural networks trained to mimic the brain's statistical properties.利用模仿大脑统计特性的神经网络来提高物体识别能力。
Neural Netw. 2020 Nov;131:103-114. doi: 10.1016/j.neunet.2020.07.013. Epub 2020 Jul 29.
6
Local features and global shape information in object classification by deep convolutional neural networks.深度卷积神经网络在目标分类中的局部特征和全局形状信息。
Vision Res. 2020 Jul;172:46-61. doi: 10.1016/j.visres.2020.04.003. Epub 2020 May 12.
7
Learning online visual invariances for novel objects via supervised and self-supervised training.通过有监督和自监督训练学习新颖对象的在线视觉不变性。
Neural Netw. 2022 Jun;150:222-236. doi: 10.1016/j.neunet.2022.02.017. Epub 2022 Mar 5.
8
Large-Scale, High-Resolution Comparison of the Core Visual Object Recognition Behavior of Humans, Monkeys, and State-of-the-Art Deep Artificial Neural Networks.大规模、高分辨率的人类、猴子和最先进的深度人工神经网络核心视觉对象识别行为比较。
J Neurosci. 2018 Aug 15;38(33):7255-7269. doi: 10.1523/JNEUROSCI.0388-18.2018. Epub 2018 Jul 13.
9
Tree-CNN: A hierarchical Deep Convolutional Neural Network for incremental learning.树卷积神经网络:一种用于增量学习的层次化深度卷积神经网络。
Neural Netw. 2020 Jan;121:148-160. doi: 10.1016/j.neunet.2019.09.010. Epub 2019 Sep 19.
10
Convolutional neural network-based encoding and decoding of visual object recognition in space and time.基于卷积神经网络的视觉目标在空间和时间上的识别的编解码。
Neuroimage. 2018 Oct 15;180(Pt A):253-266. doi: 10.1016/j.neuroimage.2017.07.018. Epub 2017 Jul 16.

引用本文的文献

1
Understanding transformation tolerant visual object representations in the human brain and convolutional neural networks.理解人类大脑和卷积神经网络中对变换具有容忍度的视觉对象表示。
Neuroimage. 2022 Nov;263:119635. doi: 10.1016/j.neuroimage.2022.119635. Epub 2022 Sep 15.

本文引用的文献

1
Scale and translation-invariance for novel objects in human vision.人类视觉中新奇物体的尺度和不变性。
Sci Rep. 2020 Jan 29;10(1):1411. doi: 10.1038/s41598-019-57261-6.
2
Deep Learning: The Good, the Bad, and the Ugly.深度学习:好的、坏的和丑的。
Annu Rev Vis Sci. 2019 Sep 15;5:399-426. doi: 10.1146/annurev-vision-091718-014951. Epub 2019 Aug 8.
3
Bayesian inference for psychology. Part II: Example applications with JASP.贝叶斯推断在心理学中的应用。第二部分:使用 JASP 的实例应用。
Psychon Bull Rev. 2018 Feb;25(1):58-76. doi: 10.3758/s13423-017-1323-7.
4
Early differential sensitivity of evoked-potentials to local and global shape during the perception of three-dimensional objects.三维物体感知过程中诱发电位对局部和整体形状的早期差异敏感性。
Neuropsychologia. 2016 Aug;89:495-509. doi: 10.1016/j.neuropsychologia.2016.07.006. Epub 2016 Jul 7.
5
The visual system supports online translation invariance for object identification.视觉系统支持用于物体识别的在线平移不变性。
Psychon Bull Rev. 2016 Apr;23(2):432-8. doi: 10.3758/s13423-015-0916-2.
6
Performance-optimized hierarchical models predict neural responses in higher visual cortex.性能优化的层次模型预测高级视觉皮层中的神经反应。
Proc Natl Acad Sci U S A. 2014 Jun 10;111(23):8619-24. doi: 10.1073/pnas.1403112111. Epub 2014 May 8.
7
Deep generative learning of location-invariant visual word recognition.位置不变视觉词识别的深度生成学习。
Front Psychol. 2013 Sep 19;4:635. doi: 10.3389/fpsyg.2013.00635. eCollection 2013.
8
Peripheral vision and pattern recognition: a review.周边视觉与模式识别:综述
J Vis. 2011 Dec 1;11(5):13. doi: 10.1167/11.5.13.
9
Complete transfer of perceptual learning across retinal locations enabled by double training.通过双重训练实现跨视网膜位置的感知学习的完全转移。
Curr Biol. 2008 Dec 23;18(24):1922-6. doi: 10.1016/j.cub.2008.10.030. Epub 2008 Dec 8.
10
Does learned shape selectivity in inferior temporal cortex automatically generalize across retinal position?颞下皮质中习得的形状选择性是否会自动在视网膜位置上泛化?
J Neurosci. 2008 Oct 1;28(40):10045-55. doi: 10.1523/JNEUROSCI.2142-08.2008.