• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

一种用于多模态分布模拟的轮廓随机梯度朗之万动力学算法。

A Contour Stochastic Gradient Langevin Dynamics Algorithm for Simulations of Multi-modal Distributions.

作者信息

Deng Wei, Lin Guang, Liang Faming

机构信息

Department of Mathematics, Purdue University, West Lafayette, IN, USA.

Departments of Mathematics & School of Mechanical Engineering, Purdue University, West Lafayette, IN, USA.

出版信息

Adv Neural Inf Process Syst. 2020 Dec;34:15725-15736.

PMID:34556969
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8457681/
Abstract

We propose an adaptively weighted stochastic gradient Langevin dynamics algorithm (SGLD), so-called contour stochastic gradient Langevin dynamics (CSGLD), for Bayesian learning in big data statistics. The proposed algorithm is essentially a , which automatically the target distribution such that the simulation for a multi-modal distribution can be greatly facilitated. Theoretically, we prove a stability condition and establish the asymptotic convergence of the self-adapting parameter to a , regardless of the non-convexity of the original energy function; we also present an error analysis for the weighted averaging estimators. Empirically, the CSGLD algorithm is tested on multiple benchmark datasets including CIFAR10 and CIFAR100. The numerical results indicate its superiority over the existing state-of-the-art algorithms in training deep neural networks.

摘要

我们提出了一种自适应加权随机梯度朗之万动力学算法(SGLD),即所谓的轮廓随机梯度朗之万动力学(CSGLD),用于大数据统计中的贝叶斯学习。所提出的算法本质上是一种,它能自动目标分布,从而极大地促进对多模态分布的模拟。从理论上讲,我们证明了一个稳定性条件,并建立了自适应参数到一个的渐近收敛性,而不考虑原始能量函数的非凸性;我们还对加权平均估计器进行了误差分析。在经验上,CSGLD算法在包括CIFAR10和CIFAR100在内的多个基准数据集上进行了测试。数值结果表明,在训练深度神经网络方面,它优于现有的最先进算法。

相似文献

1
A Contour Stochastic Gradient Langevin Dynamics Algorithm for Simulations of Multi-modal Distributions.一种用于多模态分布模拟的轮廓随机梯度朗之万动力学算法。
Adv Neural Inf Process Syst. 2020 Dec;34:15725-15736.
2
Stochastic gradient Langevin dynamics with adaptive drifts.具有自适应漂移的随机梯度朗之万动力学
J Stat Comput Simul. 2022;92(2):318-336. doi: 10.1080/00949655.2021.1958812. Epub 2021 Jul 27.
3
Non-convex Learning via Replica Exchange Stochastic Gradient MCMC.通过复制交换随机梯度马尔可夫链蒙特卡罗方法进行非凸学习
Proc Mach Learn Res. 2020 Jul;119:2474-2483.
4
Extended Stochastic Gradient MCMC for Large-Scale Bayesian Variable Selection.用于大规模贝叶斯变量选择的扩展随机梯度马尔可夫链蒙特卡罗方法
Biometrika. 2020 Dec;107(4):997-1004. doi: 10.1093/biomet/asaa029. Epub 2020 Jul 13.
5
Accelerating DNN Training Through Selective Localized Learning.通过选择性局部学习加速深度神经网络训练
Front Neurosci. 2022 Jan 11;15:759807. doi: 10.3389/fnins.2021.759807. eCollection 2021.
6
PID Controller-Based Stochastic Optimization Acceleration for Deep Neural Networks.基于 PID 控制器的深度神经网络随机优化加速。
IEEE Trans Neural Netw Learn Syst. 2020 Dec;31(12):5079-5091. doi: 10.1109/TNNLS.2019.2963066. Epub 2020 Nov 30.
7
Variance Reduction in Stochastic Gradient Langevin Dynamics.随机梯度朗之万动力学中的方差缩减
Adv Neural Inf Process Syst. 2016 Dec;29:1154-1162.
8
Stochastic Gradient Langevin dynamics for joint parameterization of tracer kinetic models, input functions, and T1 relaxation-times from undersampled k-space DCE-MRI.基于欠采样 k 空间 DCE-MRI 的示踪剂动力学模型、输入函数和 T1 弛豫时间的联合参数化的随机梯度 Langevin 动力学。
Med Image Anal. 2020 May;62:101690. doi: 10.1016/j.media.2020.101690. Epub 2020 Mar 16.
9
Geometry-informed irreversible perturbations for accelerated convergence of Langevin dynamics.用于加速朗之万动力学收敛的几何信息不可逆扰动。
Stat Comput. 2022;32(5):78. doi: 10.1007/s11222-022-10147-6. Epub 2022 Sep 19.
10
Scalable estimation strategies based on stochastic approximations: Classical results and new insights.基于随机近似的可扩展估计策略:经典结果与新见解。
Stat Comput. 2015 Jul 1;25(4):781-795. doi: 10.1007/s11222-015-9560-y.

引用本文的文献

1
A Stochastic Approximation-Langevinized Ensemble Kalman Filter Algorithm for State Space Models with Unknown Parameters.一种用于具有未知参数的状态空间模型的随机近似-朗之万化集合卡尔曼滤波算法
J Comput Graph Stat. 2023;32(2):448-469. doi: 10.1080/10618600.2022.2107531. Epub 2022 Oct 7.
2
Stochastic gradient Langevin dynamics with adaptive drifts.具有自适应漂移的随机梯度朗之万动力学
J Stat Comput Simul. 2022;92(2):318-336. doi: 10.1080/00949655.2021.1958812. Epub 2021 Jul 27.

本文引用的文献

1
Non-convex Learning via Replica Exchange Stochastic Gradient MCMC.通过复制交换随机梯度马尔可夫链蒙特卡罗方法进行非凸学习
Proc Mach Learn Res. 2020 Jul;119:2474-2483.
2
Optimization by simulated annealing.模拟退火优化。
Science. 1983 May 13;220(4598):671-80. doi: 10.1126/science.220.4598.671.
3
Efficient, multiple-range random walk algorithm to calculate the density of states.用于计算态密度的高效多范围随机游走算法。
Phys Rev Lett. 2001 Mar 5;86(10):2050-3. doi: 10.1103/PhysRevLett.86.2050.
4
Replica Monte Carlo simulation of spin glasses.自旋玻璃的复制蒙特卡罗模拟
Phys Rev Lett. 1986 Nov 24;57(21):2607-2609. doi: 10.1103/PhysRevLett.57.2607.