• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

元学习及其应用的信息论泛化界

Information-Theoretic Generalization Bounds for Meta-Learning and Applications.

作者信息

Jose Sharu Theresa, Simeone Osvaldo

机构信息

Department of Engineering, King's College London, London WC2R 2LS, UK.

出版信息

Entropy (Basel). 2021 Jan 19;23(1):126. doi: 10.3390/e23010126.

DOI:10.3390/e23010126
PMID:33478002
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7835863/
Abstract

Meta-learning, or "learning to learn", refers to techniques that infer an inductive bias from data corresponding to multiple related tasks with the goal of improving the sample efficiency for new, previously unobserved, tasks. A key performance measure for meta-learning is the meta-generalization gap, that is, the difference between the average loss measured on the meta-training data and on a new, randomly selected task. This paper presents novel information-theoretic upper bounds on the meta-generalization gap. Two broad classes of meta-learning algorithms are considered that use either separate within-task training and test sets, like model agnostic meta-learning (MAML), or joint within-task training and test sets, like reptile. Extending the existing work for conventional learning, an upper bound on the meta-generalization gap is derived for the former class that depends on the mutual information (MI) between the output of the meta-learning algorithm and its input meta-training data. For the latter, the derived bound includes an additional MI between the output of the per-task learning procedure and corresponding data set to capture within-task uncertainty. Tighter bounds are then developed for the two classes via novel individual task MI (ITMI) bounds. Applications of the derived bounds are finally discussed, including a broad class of noisy iterative algorithms for meta-learning.

摘要

元学习,即“学会学习”,指的是从与多个相关任务对应的数据中推断归纳偏差的技术,目的是提高针对新的、之前未观察到的任务的样本效率。元学习的一个关键性能指标是元泛化差距,即元训练数据上测得的平均损失与新的、随机选择的任务上测得的平均损失之间的差异。本文给出了关于元泛化差距的新的信息论上界。考虑了两大类元学习算法,一类使用单独的任务内训练集和测试集,如模型无关元学习(MAML),另一类使用联合的任务内训练集和测试集,如爬行动物算法。扩展传统学习的现有工作,为前一类算法推导出了一个依赖于元学习算法输出与其输入元训练数据之间互信息(MI)的元泛化差距上界。对于后一类算法,推导的界包括任务内学习过程的输出与相应数据集之间的额外互信息,以捕捉任务内的不确定性。然后通过新的个体任务互信息(ITMI)界为这两类算法推导出更紧的界。最后讨论了推导界的应用,包括一大类用于元学习的有噪声迭代算法。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/d94863ff60b1/entropy-23-00126-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/d9c5f2c7dd3f/entropy-23-00126-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/4615887ec857/entropy-23-00126-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/776c8d0a0ee0/entropy-23-00126-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/a39f80852a5e/entropy-23-00126-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/7ad0a1f462a2/entropy-23-00126-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/ff06006ec380/entropy-23-00126-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/d94863ff60b1/entropy-23-00126-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/d9c5f2c7dd3f/entropy-23-00126-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/4615887ec857/entropy-23-00126-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/776c8d0a0ee0/entropy-23-00126-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/a39f80852a5e/entropy-23-00126-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/7ad0a1f462a2/entropy-23-00126-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/ff06006ec380/entropy-23-00126-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/235d/7835863/d94863ff60b1/entropy-23-00126-g007.jpg

相似文献

1
Information-Theoretic Generalization Bounds for Meta-Learning and Applications.元学习及其应用的信息论泛化界
Entropy (Basel). 2021 Jan 19;23(1):126. doi: 10.3390/e23010126.
2
Algorithm-Dependent Generalization Bounds for Multi-Task Learning.多任务学习的算法相关泛化界。
IEEE Trans Pattern Anal Mach Intell. 2017 Feb;39(2):227-241. doi: 10.1109/TPAMI.2016.2544314. Epub 2016 Mar 21.
3
Improved Information-Theoretic Generalization Bounds for Distributed, Federated, and Iterative Learning.分布式、联邦式和迭代学习的改进信息论泛化界
Entropy (Basel). 2022 Aug 24;24(9):1178. doi: 10.3390/e24091178.
4
Stability-Based Generalization Analysis of Distributed Learning Algorithms for Big Data.大数据分布式学习算法基于稳定性的泛化分析
IEEE Trans Neural Netw Learn Syst. 2020 Mar;31(3):801-812. doi: 10.1109/TNNLS.2019.2910188. Epub 2019 May 8.
5
Domain Adaptation for Medical Image Segmentation: A Meta-Learning Method.用于医学图像分割的域适应:一种元学习方法。
J Imaging. 2021 Feb 10;7(2):31. doi: 10.3390/jimaging7020031.
6
Generalization Bounds of Multitask Learning From Perspective of Vector-Valued Function Learning.多任务学习的泛化界:从向量值函数学习的角度。
IEEE Trans Neural Netw Learn Syst. 2021 May;32(5):1906-1919. doi: 10.1109/TNNLS.2020.2995428. Epub 2021 May 3.
7
Going Deeper, Generalizing Better: An Information-Theoretic View for Deep Learning.深入挖掘,更好地泛化:深度学习的信息论视角
IEEE Trans Neural Netw Learn Syst. 2024 Nov;35(11):16683-16695. doi: 10.1109/TNNLS.2023.3297113. Epub 2024 Oct 29.
8
Visual Perception Generalization for Vision-and-Language Navigation via Meta-Learning.通过元学习实现视觉与语言导航的视觉感知泛化
IEEE Trans Neural Netw Learn Syst. 2023 Aug;34(8):5193-5199. doi: 10.1109/TNNLS.2021.3122579. Epub 2023 Aug 4.
9
Domain generalization on medical imaging classification using episodic training with task augmentation.基于任务增强的情景训练在医学影像分类中的领域泛化
Comput Biol Med. 2022 Feb;141:105144. doi: 10.1016/j.compbiomed.2021.105144. Epub 2021 Dec 24.
10
Sparse multinomial logistic regression: fast algorithms and generalization bounds.稀疏多项逻辑回归:快速算法与泛化界
IEEE Trans Pattern Anal Mach Intell. 2005 Jun;27(6):957-68. doi: 10.1109/TPAMI.2005.127.

引用本文的文献

1
Recent Progress of Artificial Intelligence Application in Polymer Materials.人工智能在高分子材料中的应用研究进展
Polymers (Basel). 2025 Jun 16;17(12):1667. doi: 10.3390/polym17121667.
2
Lossless Transformations and Excess Risk Bounds in Statistical Inference.统计推断中的无损变换与超额风险界
Entropy (Basel). 2023 Sep 28;25(10):1394. doi: 10.3390/e25101394.

本文引用的文献

1
Towards a Unified Theory of Learning and Information.迈向学习与信息的统一理论。
Entropy (Basel). 2020 Apr 13;22(4):438. doi: 10.3390/e22040438.
2
A general framework for updating belief distributions.用于更新信念分布的通用框架。
J R Stat Soc Series B Stat Methodol. 2016 Nov;78(5):1103-1130. doi: 10.1111/rssb.12158. Epub 2016 Feb 23.
3
General conditions for predictivity in learning theory.学习理论中预测性的一般条件。
Nature. 2004 Mar 25;428(6981):419-22. doi: 10.1038/nature02341.
4
Algorithmic stability and sanity-check bounds for leave-one-out cross-validation.留一法交叉验证的算法稳定性和合理性检查界限
Neural Comput. 1999 Aug 15;11(6):1427-53. doi: 10.1162/089976699300016304.