• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

在ProtoDUNE数据处理中使用GPU加速机器学习推理

Accelerating Machine Learning Inference with GPUs in ProtoDUNE Data Processing.

作者信息

Cai Tejin, Herner Kenneth, Yang Tingjun, Wang Michael, Acosta Flechas Maria, Harris Philip, Holzman Burt, Pedro Kevin, Tran Nhan

机构信息

Department of Physics and Astronomy, York University, 4700 Keele Street, Toronto, M3J 1P3 ON Canada.

Fermi National Accelerator Laboratory, Kirk Road and Pine Streets, Batavia, 60510 IL USA.

出版信息

Comput Softw Big Sci. 2023;7(1):11. doi: 10.1007/s41781-023-00101-0. Epub 2023 Oct 27.

DOI:10.1007/s41781-023-00101-0
PMID:37899771
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10611601/
Abstract

We study the performance of a cloud-based GPU-accelerated inference server to speed up event reconstruction in neutrino data batch jobs. Using detector data from the ProtoDUNE experiment and employing the standard DUNE grid job submission tools, we attempt to reprocess the data by running several thousand concurrent grid jobs, a rate we expect to be typical of current and future neutrino physics experiments. We process most of the dataset with the GPU version of our processing algorithm and the remainder with the CPU version for timing comparisons. We find that a 100-GPU cloud-based server is able to easily meet the processing demand, and that using the GPU version of the event processing algorithm is two times faster than processing these data with the CPU version when comparing to the newest CPUs in our sample. The amount of data transferred to the inference server during the GPU runs can overwhelm even the highest-bandwidth network switches, however, unless care is taken to observe network facility limits or otherwise distribute the jobs to multiple sites. We discuss the lessons learned from this processing campaign and several avenues for future improvements.

摘要

我们研究了基于云的GPU加速推理服务器在中微子数据批处理作业中加速事件重建的性能。利用来自ProtoDUNE实验的探测器数据,并使用标准的DUNE网格作业提交工具,我们尝试通过运行数千个并发网格作业来重新处理数据,我们预计这一速率是当前和未来中微子物理实验的典型速率。我们使用处理算法的GPU版本处理大部分数据集,其余部分使用CPU版本进行计时比较。我们发现,基于100个GPU的云服务器能够轻松满足处理需求,并且与我们样本中最新的CPU相比,使用事件处理算法的GPU版本比使用CPU版本处理这些数据快两倍。然而,在GPU运行期间传输到推理服务器的数据量甚至可能使最高带宽的网络交换机不堪重负,除非注意遵守网络设施限制或以其他方式将作业分发到多个站点。我们讨论了从这次处理活动中学到的经验教训以及未来改进的几个途径。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/2725b312171d/41781_2023_101_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/69c3a3522080/41781_2023_101_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/9bfe91356cac/41781_2023_101_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/600ffe7e0cbd/41781_2023_101_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/b4a18e4a0d65/41781_2023_101_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/5273ac9b6b75/41781_2023_101_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/1841234a5996/41781_2023_101_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/57d2b39fad92/41781_2023_101_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/2725b312171d/41781_2023_101_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/69c3a3522080/41781_2023_101_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/9bfe91356cac/41781_2023_101_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/600ffe7e0cbd/41781_2023_101_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/b4a18e4a0d65/41781_2023_101_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/5273ac9b6b75/41781_2023_101_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/1841234a5996/41781_2023_101_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/57d2b39fad92/41781_2023_101_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f30f/10611601/2725b312171d/41781_2023_101_Fig8_HTML.jpg

相似文献

1
Accelerating Machine Learning Inference with GPUs in ProtoDUNE Data Processing.在ProtoDUNE数据处理中使用GPU加速机器学习推理
Comput Softw Big Sci. 2023;7(1):11. doi: 10.1007/s41781-023-00101-0. Epub 2023 Oct 27.
2
GPU-Accelerated Machine Learning Inference as a Service for Computing in Neutrino Experiments.作为中微子实验计算服务的GPU加速机器学习推理
Front Big Data. 2021 Jan 14;3:604083. doi: 10.3389/fdata.2020.604083. eCollection 2020.
3
Multi-GPU Jacobian accelerated computing for soft-field tomography.多 GPU 雅可比加速计算在软场层析成像中的应用。
Physiol Meas. 2012 Oct;33(10):1703-15. doi: 10.1088/0967-3334/33/10/1703. Epub 2012 Sep 26.
4
Fully 3D list-mode time-of-flight PET image reconstruction on GPUs using CUDA.基于 CUDA 的 GPU 上完全 3D 列表模式飞行时间 PET 图像重建。
Med Phys. 2011 Dec;38(12):6775-86. doi: 10.1118/1.3661998.
5
Grid-based algorithm to search critical points, in the electron density, accelerated by graphics processing units.基于网格的算法,用于在电子密度中搜索临界点,由图形处理单元加速。
J Comput Chem. 2014 Dec 5;35(31):2272-8. doi: 10.1002/jcc.23752.
6
Heterogeneous computing for epidemiological model fitting and simulation.流行病学模型拟合与模拟的异构计算。
BMC Bioinformatics. 2018 Mar 16;19(1):101. doi: 10.1186/s12859-018-2108-3.
7
MIMO Radar Parallel Simulation System Based on CPU/GPU Architecture.基于 CPU/GPU 架构的 MIMO 雷达并行仿真系统。
Sensors (Basel). 2022 Jan 5;22(1):396. doi: 10.3390/s22010396.
8
Accelerating genomic workflows using NVIDIA Parabricks.利用 NVIDIA Parabricks 加速基因组工作流程。
BMC Bioinformatics. 2023 May 31;24(1):221. doi: 10.1186/s12859-023-05292-2.
9
Accelerating Spaceborne SAR Imaging Using Multiple CPU/GPU Deep Collaborative Computing.利用多个CPU/GPU深度协同计算加速星载合成孔径雷达成像
Sensors (Basel). 2016 Apr 7;16(4):494. doi: 10.3390/s16040494.
10
Efficient methods for implementation of multi-level nonrigid mass-preserving image registration on GPUs and multi-threaded CPUs.在图形处理器(GPU)和多线程中央处理器(CPU)上实现多级非刚性质量守恒图像配准的高效方法。
Comput Methods Programs Biomed. 2016 Apr;127:290-300. doi: 10.1016/j.cmpb.2015.12.018. Epub 2016 Jan 6.

引用本文的文献

1
Elucidating the role of stacking faults in TlGaSe on its thermoelectric properties.阐明层错在TlGaSe中对其热电性能的作用。
NPJ 2D Mater Appl. 2025;9(1):46. doi: 10.1038/s41699-025-00569-x. Epub 2025 Jun 6.
2
Portable Acceleration of CMS Computing Workflows with Coprocessors as a Service.通过将协处理器作为服务实现CMS计算工作流的便携式加速。
Comput Softw Big Sci. 2024;8(1):17. doi: 10.1007/s41781-024-00124-1. Epub 2024 Sep 4.

本文引用的文献

1
GPU-Accelerated Machine Learning Inference as a Service for Computing in Neutrino Experiments.作为中微子实验计算服务的GPU加速机器学习推理
Front Big Data. 2021 Jan 14;3:604083. doi: 10.3389/fdata.2020.604083. eCollection 2020.
2
The Pandora multi-algorithm approach to automated pattern recognition of cosmic-ray muon and neutrino events in the MicroBooNE detector.潘多拉多算法方法用于在MicroBooNE探测器中对宇宙射线μ子和中微子事件进行自动模式识别。
Eur Phys J C Part Fields. 2018;78(1):82. doi: 10.1140/epjc/s10052-017-5481-6. Epub 2018 Jan 29.