• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

条件熵瓶颈

The Conditional Entropy Bottleneck.

作者信息

Fischer Ian

机构信息

Google Research, Mountain View, CA 94043, USA.

出版信息

Entropy (Basel). 2020 Sep 8;22(9):999. doi: 10.3390/e22090999.

DOI:10.3390/e22090999
PMID:33286768
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7597329/
Abstract

Much of the field of Machine Learning exhibits a prominent set of failure modes, including vulnerability to adversarial examples, poor out-of-distribution (OoD) detection, miscalibration, and willingness to memorize random labelings of datasets. We characterize these as failures of , which extends the traditional measure of generalization as accuracy or related metrics on a held-out set. We hypothesize that these failures to robustly generalize are due to the learning systems retaining information about the training data. To test this hypothesis, we propose the (MNI) criterion for evaluating the quality of a model. In order to train models that perform well with respect to the MNI criterion, we present a new objective function, the (CEB), which is closely related to the (IB). We experimentally test our hypothesis by comparing the performance of CEB models with deterministic models and Variational Information Bottleneck (VIB) models on a variety of different datasets and robustness challenges. We find strong empirical evidence supporting our hypothesis that MNI models improve on these problems of robust generalization.

摘要

机器学习领域的许多方面都呈现出一系列突出的失败模式,包括易受对抗样本攻击、分布外(OoD)检测能力差、校准错误以及倾向于记住数据集的随机标注。我们将这些特征化为 的失败,它扩展了传统的泛化度量,即作为在留出集上的准确率或相关指标。我们假设这些未能稳健泛化的情况是由于学习系统保留了有关训练数据的信息。为了检验这一假设,我们提出了用于评估模型质量的 (MNI)准则。为了训练在MNI准则方面表现良好的模型,我们提出了一个新的目标函数,即 (CEB),它与 (IB)密切相关。我们通过在各种不同的数据集和鲁棒性挑战上比较CEB模型与确定性模型以及变分信息瓶颈(VIB)模型的性能,对我们的假设进行了实验测试。我们发现有力的实证证据支持我们的假设,即MNI模型在这些稳健泛化问题上有所改进。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/3552861bed16/entropy-22-00999-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/2ed38cce3c55/entropy-22-00999-g0A1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/08f8c40e8787/entropy-22-00999-g0A2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/06391838f394/entropy-22-00999-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/34a7a0770edd/entropy-22-00999-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/ef0821fb2bd8/entropy-22-00999-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/79409cc82eab/entropy-22-00999-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/b4f13b0ea8fb/entropy-22-00999-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/3552861bed16/entropy-22-00999-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/2ed38cce3c55/entropy-22-00999-g0A1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/08f8c40e8787/entropy-22-00999-g0A2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/06391838f394/entropy-22-00999-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/34a7a0770edd/entropy-22-00999-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/ef0821fb2bd8/entropy-22-00999-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/79409cc82eab/entropy-22-00999-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/b4f13b0ea8fb/entropy-22-00999-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8d5e/7597329/3552861bed16/entropy-22-00999-g006.jpg

相似文献

1
The Conditional Entropy Bottleneck.条件熵瓶颈
Entropy (Basel). 2020 Sep 8;22(9):999. doi: 10.3390/e22090999.
2
A Comparison of Variational Bounds for the Information Bottleneck Functional.信息瓶颈泛函变分界的比较
Entropy (Basel). 2020 Oct 29;22(11):1229. doi: 10.3390/e22111229.
3
Counterfactual Supervision-Based Information Bottleneck for Out-of-Distribution Generalization.基于反事实监督的信息瓶颈用于分布外泛化
Entropy (Basel). 2023 Jan 18;25(2):193. doi: 10.3390/e25020193.
4
CEB Improves Model Robustness.CEB提高模型鲁棒性。
Entropy (Basel). 2020 Sep 25;22(10):1081. doi: 10.3390/e22101081.
5
InfoAT: Improving Adversarial Training Using the Information Bottleneck Principle.InfoAT:利用信息瓶颈原理改进对抗训练
IEEE Trans Neural Netw Learn Syst. 2022 Jun 22;PP. doi: 10.1109/TNNLS.2022.3183095.
6
Adversarial Information Bottleneck.对抗性信息瓶颈
IEEE Trans Neural Netw Learn Syst. 2022 May 20;PP. doi: 10.1109/TNNLS.2022.3172986.
7
On Neural Networks Fitting, Compression, and Generalization Behavior via Information-Bottleneck-like Approaches.基于类信息瓶颈方法的神经网络拟合、压缩与泛化行为研究
Entropy (Basel). 2023 Jul 14;25(7):1063. doi: 10.3390/e25071063.
8
Nonlinear quality-related fault detection using combined deep variational information bottleneck and variational autoencoder.基于深度变分信息瓶颈与变分自编码器相结合的非线性质量相关故障检测
ISA Trans. 2021 Aug;114:444-454. doi: 10.1016/j.isatra.2021.01.002. Epub 2021 Jan 11.
9
Performance of a Computational Model of the Mammalian Olfactory System哺乳动物嗅觉系统计算模型的性能
10
Robustness to Spurious Correlations Improves Semantic Out-of-Distribution Detection.对虚假相关性的鲁棒性可改善语义分布外检测。
Proc AAAI Conf Artif Intell. 2023 Jun 27;37(12):15305-15312. doi: 10.1609/aaai.v37i12.26785.

引用本文的文献

1
Oil-Painting Style Classification Using ResNet with Conditional Information Bottleneck Regularization.使用带有条件信息瓶颈正则化的残差网络进行油画风格分类
Entropy (Basel). 2025 Jun 25;27(7):677. doi: 10.3390/e27070677.
2
The Supervised Information Bottleneck.监督信息瓶颈
Entropy (Basel). 2025 Apr 22;27(5):452. doi: 10.3390/e27050452.
3
Causal Discovery and Reasoning for Continuous Variables with an Improved Bayesian Network Constructed by Locality Sensitive Hashing and Kernel Density Estimation.基于局部敏感哈希和核密度估计构建的改进贝叶斯网络对连续变量进行因果发现与推理

本文引用的文献

1
Learning Representations for Neural Network-Based Classification Using the Information Bottleneck Principle.使用信息瓶颈原理学习基于神经网络的分类表示。
IEEE Trans Pattern Anal Mach Intell. 2020 Sep;42(9):2225-2239. doi: 10.1109/TPAMI.2019.2909031. Epub 2019 Apr 2.
2
Information Dropout: Learning Optimal Representations Through Noisy Computation.信息丢失:通过噪声计算学习最优表示
IEEE Trans Pattern Anal Mach Intell. 2018 Dec;40(12):2897-2905. doi: 10.1109/TPAMI.2017.2784440. Epub 2018 Jan 10.
3
Predictability, complexity, and learning.
Entropy (Basel). 2025 Jan 24;27(2):123. doi: 10.3390/e27020123.
4
Exploring the Trade-Off in the Variational Information Bottleneck for Regression with a Single Training Run.单次训练运行下探索回归变分信息瓶颈中的权衡
Entropy (Basel). 2024 Nov 30;26(12):1043. doi: 10.3390/e26121043.
5
Contrast Information Dynamics: A Novel Information Measure for Cognitive Modelling.对比信息动力学:一种用于认知建模的新型信息度量
Entropy (Basel). 2024 Jul 27;26(8):638. doi: 10.3390/e26080638.
6
Partial Information Decomposition: Redundancy as Information Bottleneck.部分信息分解:作为信息瓶颈的冗余度
Entropy (Basel). 2024 Jun 26;26(7):546. doi: 10.3390/e26070546.
7
The Importance of Understanding Deep Learning.理解深度学习的重要性。
Erkenntnis. 2024;89(5):1823-1840. doi: 10.1007/s10670-022-00605-y. Epub 2022 Aug 7.
8
To Compress or Not to Compress-Self-Supervised Learning and Information Theory: A Review.压缩还是不压缩——自监督学习与信息论:综述
Entropy (Basel). 2024 Mar 12;26(3):252. doi: 10.3390/e26030252.
9
High-performing neural network models of visual cortex benefit from high latent dimensionality.高表现的视觉皮层神经网络模型受益于高潜在维度。
PLoS Comput Biol. 2024 Jan 10;20(1):e1011792. doi: 10.1371/journal.pcbi.1011792. eCollection 2024 Jan.
10
Information-theoretic analysis of Hierarchical Temporal Memory-Spatial Pooler algorithm with a new upper bound for the standard information bottleneck method.基于标准信息瓶颈方法新上界的分层时间记忆-空间池化算法的信息论分析
Front Comput Neurosci. 2023 Jun 7;17:1140782. doi: 10.3389/fncom.2023.1140782. eCollection 2023.
可预测性、复杂性与学习。
Neural Comput. 2001 Nov;13(11):2409-63. doi: 10.1162/089976601753195969.