• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

不同自编码器的信息流。

Information Flows of Diverse Autoencoders.

作者信息

Lee Sungyeop, Jo Junghyo

机构信息

Department of Physics and Astronomy, Seoul National University, Seoul 08826, Korea.

Department of Physics Education and Center for Theoretical Physics and Artificial Intelligence Institute, Seoul National University, Seoul 08826, Korea.

出版信息

Entropy (Basel). 2021 Jul 5;23(7):862. doi: 10.3390/e23070862.

DOI:10.3390/e23070862
PMID:34356403
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8303402/
Abstract

Deep learning methods have had outstanding performances in various fields. A fundamental query is why they are so effective. Information theory provides a potential answer by interpreting the learning process as the information transmission and compression of data. The information flows can be visualized on the information plane of the mutual information among the input, hidden, and output layers. In this study, we examine how the information flows are shaped by the network parameters, such as depth, sparsity, weight constraints, and hidden representations. Here, we adopt autoencoders as models of deep learning, because (i) they have clear guidelines for their information flows, and (ii) they have various species, such as vanilla, sparse, tied, variational, and label autoencoders. We measured their information flows using Rényi's matrix-based α-order entropy functional. As learning progresses, they show a typical fitting phase where the amounts of input-to-hidden and hidden-to-output mutual information both increase. In the last stage of learning, however, some autoencoders show a simplifying phase, previously called the "compression phase", where input-to-hidden mutual information diminishes. In particular, the sparsity regularization of hidden activities amplifies the simplifying phase. However, tied, variational, and label autoencoders do not have a simplifying phase. Nevertheless, all autoencoders have similar reconstruction errors for training and test data. Thus, the simplifying phase does not seem to be necessary for the generalization of learning.

摘要

深度学习方法在各个领域都有着出色的表现。一个基本的问题是它们为何如此有效。信息论通过将学习过程解释为数据的信息传输和压缩提供了一个可能的答案。信息流可以在输入层、隐藏层和输出层之间的互信息的信息平面上可视化。在本研究中,我们研究了信息流是如何由网络参数塑造的,如深度、稀疏性、权重约束和隐藏表示。在这里,我们采用自动编码器作为深度学习模型,因为(i)它们的信息流有明确的指导方针,并且(ii)它们有各种类型,如普通型、稀疏型、绑定型、变分型和标签自动编码器。我们使用基于雷尼矩阵的α阶熵泛函来测量它们的信息流。随着学习的进行,它们呈现出一个典型的拟合阶段,其中输入到隐藏和隐藏到输出的互信息都增加。然而,在学习的最后阶段,一些自动编码器呈现出一个简化阶段,以前称为“压缩阶段”,其中输入到隐藏的互信息减少。特别是,隐藏活动的稀疏正则化放大了简化阶段。然而,绑定型、变分型和标签自动编码器没有简化阶段。尽管如此,所有自动编码器对于训练和测试数据都有相似的重建误差。因此,简化阶段似乎对于学习的泛化不是必需的。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/dc39261baf1b/entropy-23-00862-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/1b00632c5e0f/entropy-23-00862-g0A1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/bbe68d35839d/entropy-23-00862-g0A2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/fc073e406d97/entropy-23-00862-g0A3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/9ccb718583bf/entropy-23-00862-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/cc82292436c7/entropy-23-00862-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/6a8930f3bdea/entropy-23-00862-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/dc39261baf1b/entropy-23-00862-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/1b00632c5e0f/entropy-23-00862-g0A1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/bbe68d35839d/entropy-23-00862-g0A2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/fc073e406d97/entropy-23-00862-g0A3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/9ccb718583bf/entropy-23-00862-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/cc82292436c7/entropy-23-00862-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/6a8930f3bdea/entropy-23-00862-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/965f/8303402/dc39261baf1b/entropy-23-00862-g004.jpg

相似文献

1
Information Flows of Diverse Autoencoders.不同自编码器的信息流。
Entropy (Basel). 2021 Jul 5;23(7):862. doi: 10.3390/e23070862.
2
Understanding autoencoders with information theoretic concepts.理解基于信息论概念的自动编码器。
Neural Netw. 2019 Sep;117:104-123. doi: 10.1016/j.neunet.2019.05.003. Epub 2019 May 15.
3
Multivariate Extension of Matrix-Based Rényi's α-Order Entropy Functional.基于矩阵的雷尼α阶熵泛函的多元扩展
IEEE Trans Pattern Anal Mach Intell. 2020 Nov;42(11):2960-2966. doi: 10.1109/TPAMI.2019.2932976. Epub 2019 Aug 5.
4
Analysis of Deep Convolutional Neural Networks Using Tensor Kernels and Matrix-Based Entropy.基于张量核和矩阵熵的深度卷积神经网络分析
Entropy (Basel). 2023 Jun 3;25(6):899. doi: 10.3390/e25060899.
5
An Overview of Variational Autoencoders for Source Separation, Finance, and Bio-Signal Applications.用于源分离、金融和生物信号应用的变分自编码器概述。
Entropy (Basel). 2021 Dec 28;24(1):55. doi: 10.3390/e24010055.
6
Complex-valued autoencoders.复值自编码器。
Neural Netw. 2012 Sep;33:136-47. doi: 10.1016/j.neunet.2012.04.011. Epub 2012 May 4.
7
Convergence Behavior of DNNs with Mutual-Information-Based Regularization.基于互信息正则化的深度神经网络收敛行为
Entropy (Basel). 2020 Jun 30;22(7):727. doi: 10.3390/e22070727.
8
Unsupervised abnormality detection through mixed structure regularization (MSR) in deep sparse autoencoders.通过深度稀疏自动编码器中的混合结构正则化 (MSR) 进行无监督异常检测。
Med Phys. 2019 May;46(5):2223-2231. doi: 10.1002/mp.13464. Epub 2019 Mar 22.
9
On Neural Networks Fitting, Compression, and Generalization Behavior via Information-Bottleneck-like Approaches.基于类信息瓶颈方法的神经网络拟合、压缩与泛化行为研究
Entropy (Basel). 2023 Jul 14;25(7):1063. doi: 10.3390/e25071063.
10
Unsupervised learning of phase transitions: From principal component analysis to variational autoencoders.相变的无监督学习:从主成分分析到变分自编码器。
Phys Rev E. 2017 Aug;96(2-1):022140. doi: 10.1103/PhysRevE.96.022140. Epub 2017 Aug 18.

引用本文的文献

1
Adaptation of Autoencoder for Sparsity Reduction From Clinical Notes Representation Learning.基于自动编码器的稀疏表示学习的临床笔记自适应。
IEEE J Transl Eng Health Med. 2023 Feb 2;11:469-478. doi: 10.1109/JTEHM.2023.3241635. eCollection 2023.
2
Multivariate Time Series Information Bottleneck.多元时间序列信息瓶颈
Entropy (Basel). 2023 May 22;25(5):831. doi: 10.3390/e25050831.

本文引用的文献

1
On Information Plane Analyses of Neural Network Classifiers-A Review.神经网络分类器的信息平面分析——综述
IEEE Trans Neural Netw Learn Syst. 2022 Dec;33(12):7039-7051. doi: 10.1109/TNNLS.2021.3089037. Epub 2022 Nov 30.
2
On the Information Bottleneck Problems: Models, Connections, Applications and Information Theoretic Views.关于信息瓶颈问题:模型、联系、应用及信息论观点
Entropy (Basel). 2020 Jan 27;22(2):151. doi: 10.3390/e22020151.
3
Multivariate Extension of Matrix-Based Rényi's α-Order Entropy Functional.基于矩阵的雷尼α阶熵泛函的多元扩展
IEEE Trans Pattern Anal Mach Intell. 2020 Nov;42(11):2960-2966. doi: 10.1109/TPAMI.2019.2932976. Epub 2019 Aug 5.
4
Distributed Variational Representation Learning.分布式变分表示学习
IEEE Trans Pattern Anal Mach Intell. 2021 Jan;43(1):120-138. doi: 10.1109/TPAMI.2019.2928806. Epub 2020 Dec 4.
5
Understanding autoencoders with information theoretic concepts.理解基于信息论概念的自动编码器。
Neural Netw. 2019 Sep;117:104-123. doi: 10.1016/j.neunet.2019.05.003. Epub 2019 May 15.
6
The Potential Energy of an Autoencoder.自编码器的势能。
IEEE Trans Pattern Anal Mach Intell. 2015 Jun;37(6):1261-73. doi: 10.1109/TPAMI.2014.2362140.
7
Auto-association by multilayer perceptrons and singular value decomposition.多层感知器和奇异值分解的自联想
Biol Cybern. 1988;59(4-5):291-4. doi: 10.1007/BF00332918.