• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

混合对偶坐标上升算法(Hybrid-DCA):一种用于随机对偶坐标上升的双异步方法。

Hybrid-DCA: A double asynchronous approach for stochastic dual coordinate ascent.

作者信息

Pal Soumitra, Xu Tingyang, Yang Tianbao, Rajasekaran Sanguthevar, Bi Jinbo

机构信息

National Center for Biotechnology Information, National Library of Medicine, National Institutes of Health, Bethesda, MD 20894, USA.

Tencent AI Lab, Shenzhen, Guangzhou, 518000, China.

出版信息

J Parallel Distrib Comput. 2020 Sep;143:47-66. doi: 10.1016/j.jpdc.2020.04.002. Epub 2020 Apr 13.

DOI:10.1016/j.jpdc.2020.04.002
PMID:32699464
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7375401/
Abstract

In prior works, stochastic dual coordinate ascent (SDCA) has been parallelized in a multi-core environment where the cores communicate through shared memory, or in a multi-processor distributed memory environment where the processors communicate through message passing. In this paper, we propose a hybrid SDCA framework for multi-core clusters, the most common high performance computing environment that consists of multiple nodes each having multiple cores and its own shared memory. We distribute data across nodes where each node solves a local problem in an asynchronous parallel fashion on its cores, and then the local updates are aggregated via an asynchronous across-node update scheme. The proposed double asynchronous method converges to a global solution for -Lipschitz continuous loss functions, and at a linear convergence rate if a smooth convex loss function is used. Extensive empirical comparison has shown that our algorithm scales better than the best known shared-memory methods and runs faster than previous distributed-memory methods. Big datasets, such as one of 280 GB from the LIBSVM repository, cannot be accommodated on a single node and hence cannot be solved by a parallel algorithm. For such a dataset, our hybrid algorithm takes less than 30 seconds to achieve a duality gap of 10 on 16 nodes each using 12 cores, which is significantly faster than the best known distributed algorithms, such as CoCoA+, that take more than 160 seconds on 16 nodes.

摘要

在先前的工作中,随机对偶坐标上升法(SDCA)已在多核环境中并行化,其中各核心通过共享内存进行通信,或者在多处理器分布式内存环境中并行化,其中各处理器通过消息传递进行通信。在本文中,我们针对多核集群提出了一种混合SDCA框架,多核集群是最常见的高性能计算环境,由多个节点组成,每个节点都有多个核心及其自己的共享内存。我们将数据分布在各个节点上,每个节点在其核心上以异步并行方式解决一个局部问题,然后通过异步跨节点更新方案聚合局部更新。所提出的双重异步方法对于-Lipschitz连续损失函数收敛到全局解,并且如果使用平滑凸损失函数,则以线性收敛速率收敛。大量的实证比较表明,我们的算法比最知名的共享内存方法扩展性更好,并且比以前的分布式内存方法运行得更快。像LIBSVM存储库中280GB的数据集之一这样的大数据集无法容纳在单个节点上,因此无法通过并行算法解决。对于这样的数据集,我们的混合算法在每个使用12个核心的16个节点上实现对偶间隙为10所需的时间不到30秒,这比最知名的分布式算法(如CoCoA+)快得多,CoCoA+在16个节点上需要超过160秒。

相似文献

1
Hybrid-DCA: A double asynchronous approach for stochastic dual coordinate ascent.混合对偶坐标上升算法(Hybrid-DCA):一种用于随机对偶坐标上升的双异步方法。
J Parallel Distrib Comput. 2020 Sep;143:47-66. doi: 10.1016/j.jpdc.2020.04.002. Epub 2020 Apr 13.
2
Robust Asynchronous Stochastic Gradient-Push: Asymptotically Optimal and Network-Independent Performance for Strongly Convex Functions.鲁棒异步随机梯度推送:强凸函数的渐近最优和与网络无关的性能
J Mach Learn Res. 2020;21.
3
Linear coordinate-descent message passing for quadratic optimization.线性坐标下降消息传递的二次优化。
Neural Comput. 2012 Dec;24(12):3340-70. doi: 10.1162/NECO_a_00368. Epub 2012 Sep 12.
4
Group-Based Alternating Direction Method of Multipliers for Distributed Linear Classification.基于分组的交替方向乘子法用于分布式线性分类。
IEEE Trans Cybern. 2017 Nov;47(11):3568-3582. doi: 10.1109/TCYB.2016.2570808. Epub 2016 Jun 1.
5
Image recovery using partitioned-separable paraboloidal surrogate coordinate ascent algorithms.使用分区可分离抛物面替代坐标上升算法的图像恢复
IEEE Trans Image Process. 2002;11(3):306-17. doi: 10.1109/83.988963.
6
Asynchronous Parallel Stochastic Quasi-Newton Methods.异步并行随机拟牛顿法
Parallel Comput. 2021 Apr;101. doi: 10.1016/j.parco.2020.102721. Epub 2020 Nov 4.
7
A(DP) SGD: Asynchronous Decentralized Parallel Stochastic Gradient Descent With Differential Privacy.异步去中心化并行随机梯度下降与差分隐私。
IEEE Trans Pattern Anal Mach Intell. 2022 Nov;44(11):8036-8047. doi: 10.1109/TPAMI.2021.3107796. Epub 2022 Oct 4.
8
Parallel definition of tear film maps on distributed-memory clusters for the support of dry eye diagnosis.用于支持干眼诊断的分布式内存集群上泪膜图的并行定义。
Comput Methods Programs Biomed. 2017 Feb;139:51-60. doi: 10.1016/j.cmpb.2016.10.027. Epub 2016 Oct 31.
9
Asynchronous Parallel Large-Scale Gaussian Process Regression.异步并行大规模高斯过程回归
IEEE Trans Neural Netw Learn Syst. 2024 Jun;35(6):8683-8694. doi: 10.1109/TNNLS.2022.3200602. Epub 2024 Jun 3.
10
Reverse engineering a gene network using an asynchronous parallel evolution strategy.使用异步并行进化策略逆向工程基因网络。
BMC Syst Biol. 2010 Mar 2;4:17. doi: 10.1186/1752-0509-4-17.