• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

对比等变自监督学习改善了与灵长类动物颞下视觉区域的对齐。

Contrastive-Equivariant Self-Supervised Learning Improves Alignment with Primate Visual Area IT.

作者信息

Yerxa Thomas, Feather Jenelle, Simoncelli Eero P, Chung SueYeon

机构信息

Center for Neural Science, New York University.

Center for Computational Neuroscience, Flatiron Institute, Simons Foundation.

出版信息

Adv Neural Inf Process Syst. 2024;37:96045-96070.

PMID:40336515
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12058038/
Abstract

Models trained with self-supervised learning objectives have recently matched or surpassed models trained with traditional supervised object recognition in their ability to predict neural responses of object-selective neurons in the primate visual system. A self-supervised learning objective is arguably a more biologically plausible organizing principle, as the optimization does not require a large number of labeled examples. However, typical self-supervised objectives may result in network representations that are overly invariant to changes in the input. Here, we show that a representation with structured variability to input transformations is better aligned with known features of visual perception and neural computation. We introduce a novel framework for converting standard invariant SSL losses into "contrastive-equivariant" versions that encourage preservation of input transformations without supervised access to the transformation parameters. We demonstrate that our proposed method systematically increases the ability of models to predict responses in macaque inferior temporal cortex. Our results demonstrate the promise of incorporating known features of neural computation into task-optimization for building better models of visual cortex.

摘要

最近,使用自监督学习目标训练的模型在预测灵长类视觉系统中对象选择性神经元的神经反应能力方面,已经达到或超过了使用传统监督对象识别训练的模型。自监督学习目标可以说是一种更符合生物学原理的组织原则,因为优化过程不需要大量带标签的示例。然而,典型的自监督目标可能会导致网络表征对输入变化过度不变。在这里,我们表明,具有结构化可变性的输入变换表征与视觉感知和神经计算的已知特征更相符。我们引入了一个新颖的框架,将标准不变的自监督学习损失转换为“对比等变”版本,该版本鼓励在无监督访问变换参数的情况下保留输入变换。我们证明,我们提出的方法系统地提高了模型预测猕猴颞下皮质反应的能力。我们的结果表明,将神经计算的已知特征纳入任务优化以构建更好的视觉皮层模型具有前景。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/57a71e367169/nihms-2074495-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/cb8d3433b459/nihms-2074495-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/f64b89eb59b9/nihms-2074495-f0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/9be6d1ae8613/nihms-2074495-f0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/b7daf3659283/nihms-2074495-f0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/4d4874591ceb/nihms-2074495-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/5e888cde0666/nihms-2074495-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/5dcf9492bca2/nihms-2074495-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/57a71e367169/nihms-2074495-f0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/cb8d3433b459/nihms-2074495-f0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/f64b89eb59b9/nihms-2074495-f0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/9be6d1ae8613/nihms-2074495-f0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/b7daf3659283/nihms-2074495-f0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/4d4874591ceb/nihms-2074495-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/5e888cde0666/nihms-2074495-f0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/5dcf9492bca2/nihms-2074495-f0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3482/12058038/57a71e367169/nihms-2074495-f0004.jpg

相似文献

1
Contrastive-Equivariant Self-Supervised Learning Improves Alignment with Primate Visual Area IT.对比等变自监督学习改善了与灵长类动物颞下视觉区域的对齐。
Adv Neural Inf Process Syst. 2024;37:96045-96070.
2
Mutual Information Driven Equivariant Contrastive Learning for 3D Action Representation Learning.用于3D动作表示学习的互信息驱动等变对比学习
IEEE Trans Image Process. 2024;33:1883-1897. doi: 10.1109/TIP.2024.3372451. Epub 2024 Mar 12.
3
A Visual Encoding Model Based on Contrastive Self-Supervised Learning for Human Brain Activity along the Ventral Visual Stream.基于对比自监督学习的人类腹侧视觉通路脑活动视觉编码模型
Brain Sci. 2021 Jul 29;11(8):1004. doi: 10.3390/brainsci11081004.
4
Contrastive Representation Learning for Gaze Estimation.用于注视估计的对比表示学习
Proc Mach Learn Res. 2023;210:37-49.
5
Learning Generalized Transformation Equivariant Representations Via AutoEncoding Transformations.通过自动编码变换学习广义变换等变表示。
IEEE Trans Pattern Anal Mach Intell. 2022 Apr;44(4):2045-2057. doi: 10.1109/TPAMI.2020.3029801. Epub 2022 Mar 4.
6
Boundary-aware information maximization for self-supervised medical image segmentation.用于自监督医学图像分割的边界感知信息最大化
Med Image Anal. 2024 May;94:103150. doi: 10.1016/j.media.2024.103150. Epub 2024 Mar 28.
7
Transformation-invariant visual representations in self-organizing spiking neural networks.自组织脉冲神经网络中的变换不变视觉表示。
Front Comput Neurosci. 2012 Jul 25;6:46. doi: 10.3389/fncom.2012.00046. eCollection 2012.
8
Local contrastive loss with pseudo-label based self-training for semi-supervised medical image segmentation.基于伪标签自训练的局部对比损失的半监督医学图像分割。
Med Image Anal. 2023 Jul;87:102792. doi: 10.1016/j.media.2023.102792. Epub 2023 Mar 11.
9
A Generic Self-Supervised Framework of Learning Invariant Discriminative Features.一种学习不变判别特征的通用自监督框架。
IEEE Trans Neural Netw Learn Syst. 2024 Sep;35(9):12938-12952. doi: 10.1109/TNNLS.2023.3265607. Epub 2024 Sep 3.
10
Semi-supervised abdominal multi-organ segmentation by object-redrawing.通过对象重绘实现半监督腹部多器官分割
Med Phys. 2024 Nov;51(11):8334-8347. doi: 10.1002/mp.17364. Epub 2024 Aug 21.

本文引用的文献

1
A large-scale examination of inductive biases shaping high-level visual representation in brains and machines.大规模考察在大脑和机器中塑造高级视觉表示的归纳偏差。
Nat Commun. 2024 Oct 30;15(1):9383. doi: 10.1038/s41467-024-53147-y.
2
Factorized visual representations in the primate visual system and deep neural networks.灵长类视觉系统和深度神经网络中的因子化视觉表示。
Elife. 2024 Jul 5;13:RP91685. doi: 10.7554/eLife.91685.
3
Many but not all deep neural network audio models capture brain responses and exhibit correspondence between model stages and brain regions.
许多(但不是全部)深度神经网络音频模型可以捕捉大脑反应,并在模型阶段和大脑区域之间表现出对应关系。
PLoS Biol. 2023 Dec 13;21(12):e3002366. doi: 10.1371/journal.pbio.3002366. eCollection 2023 Dec.
4
Model metamers reveal divergent invariances between biological and artificial neural networks.模型同型揭示了生物神经网络和人工神经网络之间的不同不变性。
Nat Neurosci. 2023 Nov;26(11):2017-2034. doi: 10.1038/s41593-023-01442-0. Epub 2023 Oct 16.
5
Linear Classification of Neural Manifolds with Correlated Variability.具有相关变异性的神经流形的线性分类
Phys Rev Lett. 2023 Jul 14;131(2):027301. doi: 10.1103/PhysRevLett.131.027301.
6
A self-supervised domain-general learning framework for human ventral stream representation.一种用于人类腹侧流表示的自监督领域泛化学习框架。
Nat Commun. 2022 Jan 25;13(1):491. doi: 10.1038/s41467-022-28091-4.
7
Primary visual cortex straightens natural video trajectories.初级视皮层使自然视频轨迹变直。
Nat Commun. 2021 Oct 13;12(1):5982. doi: 10.1038/s41467-021-25939-z.
8
Unsupervised neural network models of the ventral visual stream.腹侧视觉流的无监督神经网络模型。
Proc Natl Acad Sci U S A. 2021 Jan 19;118(3). doi: 10.1073/pnas.2014196118.
9
Integrative Benchmarking to Advance Neurally Mechanistic Models of Human Intelligence.综合基准测试以推进人类智能的神经机制模型。
Neuron. 2020 Nov 11;108(3):413-423. doi: 10.1016/j.neuron.2020.07.040. Epub 2020 Sep 11.
10
Convolutional Neural Networks as a Model of the Visual System: Past, Present, and Future.作为视觉系统模型的卷积神经网络:过去、现在与未来。
J Cogn Neurosci. 2021 Sep 1;33(10):2017-2031. doi: 10.1162/jocn_a_01544.