• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

深度卷积神经网络中目标导向注意力的成本与收益

The Costs and Benefits of Goal-Directed Attention in Deep Convolutional Neural Networks.

作者信息

Luo Xiaoliang, Roads Brett D, Love Bradley C

机构信息

Department of Experimental Psychology, University College London, 26 Bedford Way, London, WC1H 0AP UK.

出版信息

Comput Brain Behav. 2021;4(2):213-230. doi: 10.1007/s42113-021-00098-y. Epub 2021 Feb 12.

DOI:10.1007/s42113-021-00098-y
PMID:34723095
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8550459/
Abstract

People deploy top-down, goal-directed attention to accomplish tasks, such as finding lost keys. By tuning the visual system to relevant information sources, object recognition can become more efficient (a benefit) and more biased toward the target (a potential cost). Motivated by selective attention in categorisation models, we developed a goal-directed attention mechanism that can process naturalistic (photographic) stimuli. Our attention mechanism can be incorporated into any existing deep convolutional neural networks (DCNNs). The processing stages in DCNNs have been related to ventral visual stream. In that light, our attentional mechanism incorporates top-down influences from prefrontal cortex (PFC) to support goal-directed behaviour. Akin to how attention weights in categorisation models warp representational spaces, we introduce a layer of attention weights to the mid-level of a DCNN that amplify or attenuate activity to further a goal. We evaluated the attentional mechanism using photographic stimuli, varying the attentional target. We found that increasing goal-directed attention has benefits (increasing hit rates) and costs (increasing false alarm rates). At a moderate level, attention improves sensitivity (i.e. increases ) at only a moderate increase in bias for tasks involving standard images, blended images and natural adversarial images chosen to fool DCNNs. These results suggest that goal-directed attention can reconfigure general-purpose DCNNs to better suit the current task goal, much like PFC modulates activity along the ventral stream. In addition to being more parsimonious and brain consistent, the mid-level attention approach performed better than a standard machine learning approach for transfer learning, namely retraining the final network layer to accommodate the new task.

摘要

人们运用自上而下的、目标导向的注意力来完成任务,比如寻找丢失的钥匙。通过将视觉系统调整到相关信息源,物体识别可以变得更高效(一个好处),同时对目标的偏向性更强(一个潜在代价)。受分类模型中选择性注意力的启发,我们开发了一种能够处理自然主义(照片)刺激的目标导向注意力机制。我们的注意力机制可以被整合到任何现有的深度卷积神经网络(DCNN)中。DCNN中的处理阶段与腹侧视觉通路有关。鉴于此,我们的注意力机制纳入了来自前额叶皮层(PFC)的自上而下的影响,以支持目标导向行为。类似于分类模型中的注意力权重如何扭曲表征空间,我们在DCNN的中间层引入了一层注意力权重,以增强或减弱活动来推进目标。我们使用照片刺激评估了注意力机制,并改变了注意力目标。我们发现增加目标导向注意力有好处(提高命中率)也有代价(提高误报率)。在适度水平上,对于涉及标准图像、混合图像和用于欺骗DCNN的自然对抗图像的任务,注意力在仅适度增加偏向性的情况下提高了敏感度(即增加了 )。这些结果表明,目标导向注意力可以重新配置通用的DCNN,使其更好地适应当前任务目标,就像PFC调节腹侧通路的活动一样。除了更简洁且与大脑一致外,中间层注意力方法在迁移学习方面比标准机器学习方法(即重新训练最终网络层以适应新任务)表现更好。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/f1f9516ed117/42113_2021_98_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/68df369e53cc/42113_2021_98_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/1f029f66399e/42113_2021_98_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/097a4c08ee3c/42113_2021_98_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/f7c1b8783278/42113_2021_98_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/49e313572e9e/42113_2021_98_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/ad74a2bcf3bf/42113_2021_98_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/f1f9516ed117/42113_2021_98_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/68df369e53cc/42113_2021_98_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/1f029f66399e/42113_2021_98_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/097a4c08ee3c/42113_2021_98_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/f7c1b8783278/42113_2021_98_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/49e313572e9e/42113_2021_98_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/ad74a2bcf3bf/42113_2021_98_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d190/8550459/f1f9516ed117/42113_2021_98_Fig10_HTML.jpg

相似文献

1
The Costs and Benefits of Goal-Directed Attention in Deep Convolutional Neural Networks.深度卷积神经网络中目标导向注意力的成本与收益
Comput Brain Behav. 2021;4(2):213-230. doi: 10.1007/s42113-021-00098-y. Epub 2021 Feb 12.
2
Correspondence between Monkey Visual Cortices and Layers of a Saliency Map Model Based on a Deep Convolutional Neural Network for Representations of Natural Images.基于深度卷积神经网络的自然图像表示的显著性映射模型的猴子视觉皮层与层之间的对应关系。
eNeuro. 2021 Feb 9;8(1). doi: 10.1523/ENEURO.0200-20.2020. Print 2021 Jan-Feb.
3
Improved object recognition using neural networks trained to mimic the brain's statistical properties.利用模仿大脑统计特性的神经网络来提高物体识别能力。
Neural Netw. 2020 Nov;131:103-114. doi: 10.1016/j.neunet.2020.07.013. Epub 2020 Jul 29.
4
Human Visual Cortex and Deep Convolutional Neural Network Care Deeply about Object Background.人类视觉皮层和深度卷积神经网络非常关注物体背景。
J Cogn Neurosci. 2024 Mar 1;36(3):551-566. doi: 10.1162/jocn_a_02098.
5
Modeling naturalistic face processing in humans with deep convolutional neural networks.用深度卷积神经网络对人类自然主义面孔处理进行建模。
Proc Natl Acad Sci U S A. 2023 Oct 24;120(43):e2304085120. doi: 10.1073/pnas.2304085120. Epub 2023 Oct 17.
6
Humans and Deep Networks Largely Agree on Which Kinds of Variation Make Object Recognition Harder.人类和深度网络在哪些类型的变化会使目标识别更困难这一问题上基本达成一致。
Front Comput Neurosci. 2016 Aug 31;10:92. doi: 10.3389/fncom.2016.00092. eCollection 2016.
7
Bottom-Up and Top-Down Factors Differentially Influence Stimulus Representations Across Large-Scale Attentional Networks.自底向上和自顶向下因素对大规模注意网络中的刺激表现有不同影响。
J Neurosci. 2018 Mar 7;38(10):2495-2504. doi: 10.1523/JNEUROSCI.2724-17.2018. Epub 2018 Feb 2.
8
Face Recognition Depends on Specialized Mechanisms Tuned to View-Invariant Facial Features: Insights from Deep Neural Networks Optimized for Face or Object Recognition.人脸识别依赖于专门的机制,这些机制针对的是不变的面部特征:来自专门针对人脸或物体识别进行优化的深度神经网络的见解。
Cogn Sci. 2021 Sep;45(9):e13031. doi: 10.1111/cogs.13031.
9
Goal-directed attention alters the tuning of object-based representations in extrastriate cortex.目标导向性注意会改变纹外皮层中基于客体的表征的调谐。
Front Hum Neurosci. 2012 Jun 21;6:187. doi: 10.3389/fnhum.2012.00187. eCollection 2012.
10
Trial history effects in the ventral attentional network.腹侧注意网络中的试验历史效应。
J Cogn Neurosci. 2014 Dec;26(12):2789-97. doi: 10.1162/jocn_a_00678. Epub 2014 Jun 24.

引用本文的文献

1
Coordinating multiple mental faculties during learning.学习过程中协调多种心理能力。
Sci Rep. 2025 Feb 13;15(1):5319. doi: 10.1038/s41598-025-89732-4.
2
Object-based feedback attention in convolutional neural networks improves tumour detection in digital pathology.卷积神经网络中基于对象的反馈注意力机制可改善数字病理学中的肿瘤检测。
Sci Rep. 2024 Dec 5;14(1):30400. doi: 10.1038/s41598-024-80717-3.
3
Tracking the Emergence of Location-based Spatial Representations in Human Scene-Selective Cortex.追踪基于位置的空间表象在人类场景选择性大脑皮层中的出现。

本文引用的文献

1
Ventromedial prefrontal cortex compression during concept learning.在概念学习期间对腹内侧前额叶皮质进行压缩。
Nat Commun. 2020 Jan 7;11(1):46. doi: 10.1038/s41467-019-13930-8.
2
THINGS: A database of 1,854 object concepts and more than 26,000 naturalistic object images.事物数据库:包含 1854 个物体概念和 26000 多张自然物体图像。
PLoS One. 2019 Oct 15;14(10):e0223792. doi: 10.1371/journal.pone.0223792. eCollection 2019.
3
Occipitotemporal representations reflect individual differences in conceptual knowledge.枕颞叶代表反映了概念知识的个体差异。
J Cogn Neurosci. 2021 Mar;33(3):445-462. doi: 10.1162/jocn_a_01654. Epub 2020 Dec 7.
4
Attention in Psychology, Neuroscience, and Machine Learning.心理学、神经科学和机器学习中的注意力
Front Comput Neurosci. 2020 Apr 16;14:29. doi: 10.3389/fncom.2020.00029. eCollection 2020.
J Exp Psychol Gen. 2019 Jul;148(7):1192-1203. doi: 10.1037/xge0000501. Epub 2018 Nov 1.
4
How biological attention mechanisms improve task performance in a large-scale visual system model.生物注意机制如何提高大规模视觉系统模型中的任务表现。
Elife. 2018 Oct 1;7:e38105. doi: 10.7554/eLife.38105.
5
Evaluating (and Improving) the Correspondence Between Deep Neural Networks and Human Representations.评估(和改进)深度神经网络与人的表示之间的对应关系。
Cogn Sci. 2018 Nov;42(8):2648-2669. doi: 10.1111/cogs.12670. Epub 2018 Sep 3.
6
Estimating the functional dimensionality of neural representations.估计神经表示的功能维度。
Neuroimage. 2018 Oct 1;179:51-62. doi: 10.1016/j.neuroimage.2018.06.015. Epub 2018 Jun 7.
7
Costs of Selective Attention: When Children Notice What Adults Miss.选择性注意的代价:当儿童注意到成人忽略的事物时。
Psychol Sci. 2017 Jun;28(6):723-732. doi: 10.1177/0956797617693005. Epub 2017 Apr 7.
8
Dynamic updating of hippocampal object representations reflects new conceptual knowledge.海马体物体表征的动态更新反映了新的概念性知识。
Proc Natl Acad Sci U S A. 2016 Nov 15;113(46):13203-13208. doi: 10.1073/pnas.1614048113. Epub 2016 Nov 1.
9
There's Waldo! A Normalization Model of Visual Search Predicts Single-Trial Human Fixations in an Object Search Task.找到了!视觉搜索的归一化模型可预测物体搜索任务中的单次试验人类注视点。
Cereb Cortex. 2016 Jul;26(7):3064-82. doi: 10.1093/cercor/bhv129. Epub 2015 Jun 19.
10
Bottom-up and top-down attention: different processes and overlapping neural systems.自下而上和自上而下的注意力:不同的过程与重叠的神经系统。
Neuroscientist. 2014 Oct;20(5):509-21. doi: 10.1177/1073858413514136. Epub 2013 Dec 20.