• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

贝叶斯神经网络训练的随机控制

Stochastic Control for Bayesian Neural Network Training.

作者信息

Winkler Ludwig, Ojeda César, Opper Manfred

机构信息

Machine Learning Group, Technische Universität Berlin, 10623 Berlin, Germany.

Artificial Intelligence Group, Technische Universität Berlin, 10623 Berlin, Germany.

出版信息

Entropy (Basel). 2022 Aug 9;24(8):1097. doi: 10.3390/e24081097.

DOI:10.3390/e24081097
PMID:36010761
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9407447/
Abstract

In this paper, we propose to leverage the Bayesian uncertainty information encoded in parameter distributions to inform the learning procedure for Bayesian models. We derive a first principle stochastic differential equation for the training dynamics of the mean and uncertainty parameter in the variational distributions. On the basis of the derived Bayesian stochastic differential equation, we apply the methodology of stochastic optimal control on the variational parameters to obtain individually controlled learning rates. We show that the resulting optimizer, StochControlSGD, is significantly more robust to large learning rates and can adaptively and individually control the learning rates of the variational parameters. The evolution of the control suggests separate and distinct dynamical behaviours in the training regimes for the mean and uncertainty parameters in Bayesian neural networks.

摘要

在本文中,我们提议利用参数分布中编码的贝叶斯不确定性信息来为贝叶斯模型的学习过程提供信息。我们推导了变分分布中均值和不确定性参数训练动态的第一原理随机微分方程。基于推导得到的贝叶斯随机微分方程,我们将随机最优控制方法应用于变分参数,以获得个体控制的学习率。我们表明,由此产生的优化器StochControlSGD对大学习率具有显著更强的鲁棒性,并且可以自适应地、个体地控制变分参数的学习率。控制的演变表明,贝叶斯神经网络中均值和不确定性参数的训练模式存在分离且不同的动态行为。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915b/9407447/aa9469134bb4/entropy-24-01097-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915b/9407447/375a22d1b54e/entropy-24-01097-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915b/9407447/d784608ddbe2/entropy-24-01097-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915b/9407447/392ba51adcd0/entropy-24-01097-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915b/9407447/aad6f8f86696/entropy-24-01097-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915b/9407447/aa9469134bb4/entropy-24-01097-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915b/9407447/375a22d1b54e/entropy-24-01097-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915b/9407447/d784608ddbe2/entropy-24-01097-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915b/9407447/392ba51adcd0/entropy-24-01097-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915b/9407447/aad6f8f86696/entropy-24-01097-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/915b/9407447/aa9469134bb4/entropy-24-01097-g005.jpg

相似文献

1
Stochastic Control for Bayesian Neural Network Training.贝叶斯神经网络训练的随机控制
Entropy (Basel). 2022 Aug 9;24(8):1097. doi: 10.3390/e24081097.
2
Measuring the Uncertainty of Predictions in Deep Neural Networks with Variational Inference.用变分推断测量深度神经网络的预测不确定性。
Sensors (Basel). 2020 Oct 23;20(21):6011. doi: 10.3390/s20216011.
3
Variational HyperAdam: A Meta-Learning Approach to Network Training.变分超Adam:一种网络训练的元学习方法。
IEEE Trans Pattern Anal Mach Intell. 2022 Aug;44(8):4469-4484. doi: 10.1109/TPAMI.2021.3061581. Epub 2022 Jul 1.
4
Stochastic complexities of general mixture models in variational Bayesian learning.变分贝叶斯学习中一般混合模型的随机复杂性
Neural Netw. 2007 Mar;20(2):210-9. doi: 10.1016/j.neunet.2006.05.030. Epub 2006 Aug 10.
5
The Inverse of Exact Renormalization Group Flows as Statistical Inference.作为统计推断的精确重整化群流的逆
Entropy (Basel). 2024 Apr 30;26(5):389. doi: 10.3390/e26050389.
6
Correlated Parameters to Accurately Measure Uncertainty in Deep Neural Networks.用于准确测量深度神经网络不确定性的相关参数。
IEEE Trans Neural Netw Learn Syst. 2021 Mar;32(3):1037-1051. doi: 10.1109/TNNLS.2020.2980004. Epub 2021 Mar 1.
7
General-Purpose Bayesian Tensor Learning With Automatic Rank Determination and Uncertainty Quantification.具有自动秩确定和不确定性量化的通用贝叶斯张量学习
Front Artif Intell. 2022 Jan 7;4:668353. doi: 10.3389/frai.2021.668353. eCollection 2021.
8
A stochastic variational framework for Recurrent Gaussian Processes models.循环高斯过程模型的随机变分框架。
Neural Netw. 2019 Apr;112:54-72. doi: 10.1016/j.neunet.2019.01.005. Epub 2019 Feb 1.
9
Variational Bayesian identification and prediction of stochastic nonlinear dynamic causal models.随机非线性动态因果模型的变分贝叶斯识别与预测
Physica D. 2009 Nov 1;238(21):2089-2118. doi: 10.1016/j.physd.2009.08.002.
10
Gradient Regularization as Approximate Variational Inference.作为近似变分推断的梯度正则化
Entropy (Basel). 2021 Dec 3;23(12):1629. doi: 10.3390/e23121629.

引用本文的文献

1
Synchronization of Complex Dynamical Networks with Stochastic Links Dynamics.具有随机链路动力学的复杂动态网络同步
Entropy (Basel). 2023 Oct 17;25(10):1457. doi: 10.3390/e25101457.

本文引用的文献

1
On-line learning in changing environments with applications in supervised and unsupervised learning.
Neural Netw. 2002 Jun-Jul;15(4-6):743-60. doi: 10.1016/s0893-6080(02)00060-6.