• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

绘制深度学习网络的学习曲线。

Mapping the learning curves of deep learning networks.

作者信息

Jiang Yanru, Dale Rick

机构信息

Department of Communication, University of California, Los Angeles, Los Angeles, California, United States of America.

出版信息

PLoS Comput Biol. 2025 Feb 10;21(2):e1012286. doi: 10.1371/journal.pcbi.1012286. eCollection 2025 Feb.

DOI:10.1371/journal.pcbi.1012286
PMID:39928655
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11841907/
Abstract

There is an important challenge in systematically interpreting the internal representations of deep neural networks (DNNs). Existing techniques are often less effective for non-tabular tasks, or they primarily focus on qualitative, ad-hoc interpretations of models. In response, this study introduces a cognitive science-inspired, multi-dimensional quantification and visualization approach that captures two temporal dimensions of model learning: the "information-processing trajectory" and the "developmental trajectory." The former represents the influence of incoming signals on an agent's decision-making, while the latter conceptualizes the gradual improvement in an agent's performance throughout its lifespan. Tracking the learning curves of DNNs enables researchers to explicitly identify the model appropriateness of a given task, examine the properties of the underlying input signals, and assess the model's alignment (or lack thereof) with human learning experiences. To illustrate this method, we conducted 750 runs of simulations on two temporal tasks: gesture detection and sentence classification, showcasing its applicability across different types of deep learning tasks. Using four descriptive metrics to quantify the mapped learning curves-start, end - start, max, tmax-, we identified significant differences in learning patterns based on data sources and class distinctions (all p's  <  .0001), the prominent role of spatial semantics in gesture learning, and larger information gains in language learning. We highlight three key insights gained from mapping learning curves: non-monotonic progress, pairwise comparisons, and domain distinctions. We reflect on the theoretical implications of this method for cognitive processing, language models and representations from multiple modalities.

摘要

在系统地解释深度神经网络(DNN)的内部表示方面存在一个重要挑战。现有技术对于非表格任务通常效果较差,或者它们主要侧重于对模型进行定性的、临时的解释。作为回应,本研究引入了一种受认知科学启发的多维度量化和可视化方法,该方法捕捉了模型学习的两个时间维度:“信息处理轨迹”和“发展轨迹”。前者表示传入信号对智能体决策的影响,而后者将智能体在其整个生命周期内性能的逐渐提升概念化。跟踪DNN的学习曲线使研究人员能够明确确定给定任务的模型适用性,检查底层输入信号的属性,并评估模型与人类学习经验的一致性(或不一致性)。为了说明这种方法,我们在两个时间任务上进行了750次模拟:手势检测和句子分类,展示了其在不同类型深度学习任务中的适用性。使用四个描述性指标来量化映射的学习曲线——起始点、终点 - 起始点、最大值、最大值对应的时间点——我们基于数据源和类别差异确定了学习模式的显著差异(所有p值 <.0001),空间语义在手势学习中的突出作用,以及语言学习中更大的信息增益。我们强调了从映射学习曲线中获得的三个关键见解:非单调进展、成对比较和领域差异。我们思考了这种方法对认知处理、语言模型和多模态表示的理论意义。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/79163fd30c04/pcbi.1012286.g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/ec03472d80d0/pcbi.1012286.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/9f5d12fdad98/pcbi.1012286.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/ce64c9cbbe6c/pcbi.1012286.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/042f5f41af86/pcbi.1012286.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/88d8d910762d/pcbi.1012286.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/437e285e3c47/pcbi.1012286.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/ae279b10c6ab/pcbi.1012286.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/79163fd30c04/pcbi.1012286.g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/ec03472d80d0/pcbi.1012286.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/9f5d12fdad98/pcbi.1012286.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/ce64c9cbbe6c/pcbi.1012286.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/042f5f41af86/pcbi.1012286.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/88d8d910762d/pcbi.1012286.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/437e285e3c47/pcbi.1012286.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/ae279b10c6ab/pcbi.1012286.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/78bf/11841907/79163fd30c04/pcbi.1012286.g009.jpg

相似文献

1
Mapping the learning curves of deep learning networks.绘制深度学习网络的学习曲线。
PLoS Comput Biol. 2025 Feb 10;21(2):e1012286. doi: 10.1371/journal.pcbi.1012286. eCollection 2025 Feb.
2
Brain tumor segmentation and detection in MRI using convolutional neural networks and VGG16.使用卷积神经网络和VGG16在磁共振成像(MRI)中进行脑肿瘤分割与检测
Cancer Biomark. 2025 Mar;42(3):18758592241311184. doi: 10.1177/18758592241311184. Epub 2025 Apr 4.
3
Neural Encoding and Decoding With Distributed Sentence Representations.分布式句子表示的神经编码和解码。
IEEE Trans Neural Netw Learn Syst. 2021 Feb;32(2):589-603. doi: 10.1109/TNNLS.2020.3027595. Epub 2021 Feb 4.
4
Symbolic Deep Networks: A Psychologically Inspired Lightweight and Efficient Approach to Deep Learning.符号深度学习网络:一种受心理学启发的轻量级高效深度学习方法。
Top Cogn Sci. 2022 Oct;14(4):702-717. doi: 10.1111/tops.12571. Epub 2021 Oct 5.
5
Systematic review of experimental paradigms and deep neural networks for electroencephalography-based cognitive workload detection.基于脑电图的认知负荷检测的实验范式和深度神经网络的系统综述。
Prog Biomed Eng (Bristol). 2024 Oct 21;6(4). doi: 10.1088/2516-1091/ad8530.
6
Cognitive decline assessment using semantic linguistic content and transformer deep learning architecture.使用语义语言内容和变压器深度学习架构评估认知能力下降。
Int J Lang Commun Disord. 2024 May-Jun;59(3):1110-1127. doi: 10.1111/1460-6984.12973. Epub 2023 Nov 16.
7
Deep Artificial Neural Networks Reveal a Distributed Cortical Network Encoding Propositional Sentence-Level Meaning.深度人工神经网络揭示命题句级意义的分布式皮层网络编码。
J Neurosci. 2021 May 5;41(18):4100-4119. doi: 10.1523/JNEUROSCI.1152-20.2021. Epub 2021 Mar 22.
8
Deep learning uncertainty quantification for clinical text classification.深度学习在临床文本分类中的不确定性量化。
J Biomed Inform. 2024 Jan;149:104576. doi: 10.1016/j.jbi.2023.104576. Epub 2023 Dec 13.
9
Deep learning for electroencephalogram (EEG) classification tasks: a review.深度学习在脑电图(EEG)分类任务中的应用:综述。
J Neural Eng. 2019 Jun;16(3):031001. doi: 10.1088/1741-2552/ab0ab5. Epub 2019 Feb 26.
10
Learning Spatial-Spectral-Temporal EEG Representations with Deep Attentive-Recurrent-Convolutional Neural Networks for Pain Intensity Assessment.利用深度注意-递归-卷积神经网络学习空间-谱-时 EEG 表示,用于疼痛强度评估。
Neuroscience. 2022 Jan 15;481:144-155. doi: 10.1016/j.neuroscience.2021.11.034. Epub 2021 Nov 26.

引用本文的文献

1
The Effect of Data Leakage and Feature Selection on Machine Learning Performance for Early Parkinson's Disease Detection.数据泄露和特征选择对早期帕金森病检测机器学习性能的影响
Bioengineering (Basel). 2025 Aug 6;12(8):845. doi: 10.3390/bioengineering12080845.

本文引用的文献

1
Phonemic segmentation of narrative speech in human cerebral cortex.人类大脑皮层叙事语音的音位切分。
Nat Commun. 2023 Jul 18;14(1):4309. doi: 10.1038/s41467-023-39872-w.
2
Modeling Structure-Building in the Brain With CCG Parsing and Large Language Models.用 CCG 解析和大型语言模型构建大脑的结构模型。
Cogn Sci. 2023 Jul;47(7):e13312. doi: 10.1111/cogs.13312.
3
Shared computational principles for language processing in humans and deep language models.人类和深度语言模型语言处理的共享计算原则。
Nat Neurosci. 2022 Mar;25(3):369-380. doi: 10.1038/s41593-022-01026-4. Epub 2022 Mar 7.
4
The neural architecture of language: Integrative modeling converges on predictive processing.语言的神经结构:综合建模趋向于预测处理。
Proc Natl Acad Sci U S A. 2021 Nov 9;118(45). doi: 10.1073/pnas.2105646118.
5
Individual differences among deep neural network models.深度神经网络模型的个体差异。
Nat Commun. 2020 Nov 12;11(1):5725. doi: 10.1038/s41467-020-19632-w.
6
Direct Fit to Nature: An Evolutionary Perspective on Biological and Artificial Neural Networks.直接契合自然:生物和人工神经网络的进化视角。
Neuron. 2020 Feb 5;105(3):416-434. doi: 10.1016/j.neuron.2019.12.002.
7
THINGS: A database of 1,854 object concepts and more than 26,000 naturalistic object images.事物数据库:包含 1854 个物体概念和 26000 多张自然物体图像。
PLoS One. 2019 Oct 15;14(10):e0223792. doi: 10.1371/journal.pone.0223792. eCollection 2019.
8
OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields.OpenPose:基于部件亲和力字段的实时多人 2D 姿态估计。
IEEE Trans Pattern Anal Mach Intell. 2021 Jan;43(1):172-186. doi: 10.1109/TPAMI.2019.2929257. Epub 2020 Dec 4.
9
Back to the future: The return of cognitive functionalism.回到未来:认知功能主义的回归。
Behav Brain Sci. 2017 Jan;40:e257. doi: 10.1017/S0140525X17000061.
10
Deconstructing multivariate decoding for the study of brain function.对大脑功能研究的多元解码进行解构。
Neuroimage. 2018 Oct 15;180(Pt A):4-18. doi: 10.1016/j.neuroimage.2017.08.005. Epub 2017 Aug 4.