• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

量化神经网络中数据类的可分离性。

Quantifying the separability of data classes in neural networks.

机构信息

Laboratory of Sensory and Cognitive Neuroscience, Aix-Marseille University, Marseille, France; Neuroscience Lab, University Hospital Erlangen, Germany; Cognitive Computational Neuroscience Group, University Erlangen-Nürnberg (FAU), Germany.

Chair of Machine Intelligence, University Erlangen-Nürnberg (FAU), Germany.

出版信息

Neural Netw. 2021 Jul;139:278-293. doi: 10.1016/j.neunet.2021.03.035. Epub 2021 Apr 5.

DOI:10.1016/j.neunet.2021.03.035
PMID:33862387
Abstract

We introduce the Generalized Discrimination Value (GDV) that measures, in a non-invasive manner, how well different data classes separate in each given layer of an artificial neural network. It turns out that, at the end of the training period, the GDV in each given layer L attains a highly reproducible value, irrespective of the initialization of the network's connection weights. In the case of multi-layer perceptrons trained with error backpropagation, we find that classification of highly complex data sets requires a temporal reduction of class separability, marked by a characteristic 'energy barrier' in the initial part of the GDV(L) curve. Even more surprisingly, for a given data set, the GDV(L) is running through a fixed 'master curve', independently from the total number of network layers. Finally, due to its invariance with respect to dimensionality, the GDV may serve as a useful tool to compare the internal representational dynamics of artificial neural networks with different architectures for neural architecture search or network compression; or even with brain activity in order to decide between different candidate models of brain function.

摘要

我们引入了广义判别值(GDV),它以非侵入性的方式衡量不同数据类在人工神经网络的每个给定层中分离的程度。事实证明,在训练期末,GDV 在每个给定层 L 中达到一个高度可重复的值,而与网络连接权重的初始化无关。在使用误差反向传播训练的多层感知机的情况下,我们发现高度复杂数据集的分类需要类可分离性的时间减少,这在 GDV(L)曲线的初始部分标记为特征“能量障碍”。更令人惊讶的是,对于给定的数据集,GDV(L) 正在经历一个固定的“主曲线”,与网络层数的总数无关。最后,由于其对维度的不变性,GDV 可以作为一种有用的工具,用于比较具有不同架构的人工神经网络的内部表示动态,以进行神经架构搜索或网络压缩;甚至可以与大脑活动进行比较,以在不同的候选大脑功能模型之间做出决策。

相似文献

1
Quantifying the separability of data classes in neural networks.量化神经网络中数据类的可分离性。
Neural Netw. 2021 Jul;139:278-293. doi: 10.1016/j.neunet.2021.03.035. Epub 2021 Apr 5.
2
Memristors for Neuromorphic Circuits and Artificial Intelligence Applications.用于神经形态电路和人工智能应用的忆阻器
Materials (Basel). 2020 Feb 20;13(4):938. doi: 10.3390/ma13040938.
3
Optimizing neural networks for medical data sets: A case study on neonatal apnea prediction.优化神经网络在医学数据集上的应用:以新生儿呼吸暂停预测为例的研究
Artif Intell Med. 2019 Jul;98:59-76. doi: 10.1016/j.artmed.2019.07.008. Epub 2019 Jul 25.
4
Siamese Neural Networks: An Overview.暹罗神经网络:概述。
Methods Mol Biol. 2021;2190:73-94. doi: 10.1007/978-1-0716-0826-5_3.
5
An effective SteinGLM initialization scheme for training multi-layer feedforward sigmoidal neural networks.一种用于训练多层前馈 Sigmoidal 神经网络的有效 SteinGLM 初始化方案。
Neural Netw. 2021 Jul;139:149-157. doi: 10.1016/j.neunet.2021.02.014. Epub 2021 Feb 27.
6
CiwGAN and fiwGAN: Encoding information in acoustic data to model lexical learning with Generative Adversarial Networks.CiwGAN 和 fiwGAN:利用生成对抗网络将声学数据中的信息编码,以建模词汇学习。
Neural Netw. 2021 Jul;139:305-325. doi: 10.1016/j.neunet.2021.03.017. Epub 2021 Mar 19.
7
Three learning phases for radial-basis-function networks.径向基函数网络的三个学习阶段。
Neural Netw. 2001 May;14(4-5):439-58. doi: 10.1016/s0893-6080(01)00027-2.
8
Words as a window: Using word embeddings to explore the learned representations of Convolutional Neural Networks.文字之窗:利用词嵌入探索卷积神经网络的习得表示。
Neural Netw. 2021 May;137:63-74. doi: 10.1016/j.neunet.2020.12.009. Epub 2021 Jan 22.
9
Neural networks for signal processing applications: ECG classification.用于信号处理应用的神经网络:心电图分类。
Australas Phys Eng Sci Med. 1997 Sep;20(3):147-51.
10
A learning rule for very simple universal approximators consisting of a single layer of perceptrons.一种由单层感知器组成的非常简单的通用逼近器的学习规则。
Neural Netw. 2008 Jun;21(5):786-95. doi: 10.1016/j.neunet.2007.12.036. Epub 2007 Dec 31.

引用本文的文献

1
Analysis of argument structure constructions in a deep recurrent language model.深度循环语言模型中论证结构构建的分析
Front Comput Neurosci. 2025 Jun 16;19:1474860. doi: 10.3389/fncom.2025.1474860. eCollection 2025.
2
Analysis of argument structure constructions in the large language model BERT.大型语言模型BERT中论证结构构建的分析
Front Artif Intell. 2025 Jan 31;8:1477246. doi: 10.3389/frai.2025.1477246. eCollection 2025.
3
Predicting High-Strength Concrete's Compressive Strength: A Comparative Study of Artificial Neural Networks, Adaptive Neuro-Fuzzy Inference System, and Response Surface Methodology.
预测高强混凝土的抗压强度:人工神经网络、自适应神经模糊推理系统和响应面法的比较研究
Materials (Basel). 2024 Sep 15;17(18):4533. doi: 10.3390/ma17184533.
4
Coincidence detection and integration behavior in spiking neural networks.脉冲神经网络中的巧合检测与整合行为
Cogn Neurodyn. 2024 Aug;18(4):1753-1765. doi: 10.1007/s11571-023-10038-0. Epub 2023 Dec 13.
5
Non-uniform Speaker Disentanglement For Depression Detection From Raw Speech Signals.基于原始语音信号的抑郁症检测的非均匀说话人解缠
Interspeech. 2023 Aug;2023:2343-2347. doi: 10.21437/interspeech.2023-2101.
6
Predictive coding and stochastic resonance as fundamental principles of auditory phantom perception.预测编码和随机共振作为听觉幻觉感知的基本原理。
Brain. 2023 Dec 1;146(12):4809-4825. doi: 10.1093/brain/awad255.
7
Extracting continuous sleep depth from EEG data without machine learning.无需机器学习从脑电图(EEG)数据中提取连续睡眠深度
Neurobiol Sleep Circadian Rhythms. 2023 May 19;14:100097. doi: 10.1016/j.nbscr.2023.100097. eCollection 2023 May.
8
Neural network based formation of cognitive maps of semantic spaces and the putative emergence of abstract concepts.基于神经网络的语义空间认知图的形成和抽象概念的假设出现。
Sci Rep. 2023 Mar 4;13(1):3644. doi: 10.1038/s41598-023-30307-6.
9
Accurate and fast clade assignment via deep learning and frequency chaos game representation.通过深度学习和频率混沌游戏表示实现准确快速的进化枝分配。
Gigascience. 2022 Dec 28;12. doi: 10.1093/gigascience/giac119.
10
Classification at the accuracy limit: facing the problem of data ambiguity.分类达到精度极限:面临数据歧义问题。
Sci Rep. 2022 Dec 21;12(1):22121. doi: 10.1038/s41598-022-26498-z.