• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

实现碳高效人工智能的模型压缩技术比较分析

Comparative analysis of model compression techniques for achieving carbon efficient AI.

作者信息

Paula Eileen, Soni Jayesh, Upadhyay Himanshu, Lagos Leonel

机构信息

Applied Research Center, Florida International University, Miami, 33174, USA.

Department of Electrical and Computer Engineering, Florida International University, Miami, 33174, USA.

出版信息

Sci Rep. 2025 Jul 2;15(1):23461. doi: 10.1038/s41598-025-07821-w.

DOI:10.1038/s41598-025-07821-w
PMID:40604122
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12222641/
Abstract

The growing computational demands of models, such as BERT, have raised concerns about their environmental impact. This study addresses the pressing need for sustainable Artificial Intelligence practices by investigating the efficiency of model compression techniques in reducing the energy consumption and carbon emissions of transformer-based models without compromising performance. Specifically, we applied pruning, knowledge distillation, and quantization to transformer-based models (BERT, DistilBERT, ALBERT, and ELECTRA) using the Amazon Polarity Dataset for sentiment analysis. We also compared the energy efficiency of these compressed models against inherently carbon-efficient transformer models, such as TinyBERT and MobileBERT. To evaluate each model's energy consumption and carbon emissions, we utilized the open-source tool CodeCarbon. Our findings indicate that applying model compression techniques resulted in a reduction in energy consumption of 32.097% for BERT with pruning and distillation, [Formula: see text]% for DistilBERT with pruning, 7.12% for ALBERT with quantization, and 23.934% for ELECTRA with pruning and distillation, while maintaining performance metrics within a range of 95.871-99.062% accuracy, precision, recall, F1 score, and ROC AUC except for ALBERT with quantization. Specifically, BERT with pruning and distillation achieved 95.90% accuracy, 95.90% precision, 95.90% recall, 95.90% F1-score, and 98.87% ROC AUC; DistilBERT with pruning achieved 95.87% accuracy, 95.87% precision, 95.87% recall, 95.87% F1-score, and 99.06% ROC AUC; ELECTRA with pruning and distillation achieved 95.92% accuracy, 95.92% precision, 95.92% recall, 95.92% F1-score, and 99.30% ROC AUC; and ALBERT with quantization achieved 65.44% accuracy, 67.82% precision, 65.44% recall, 63.46% F1-score, and 72.31% ROC AUC, indicating significant performance degradation due to quantization sensitivity in its already compressed architecture. Overall, this demonstrates the potential for sustainable Artificial Intelligence practices using model compression.

摘要

诸如BERT等模型不断增长的计算需求引发了人们对其环境影响的担忧。本研究通过调查模型压缩技术在不影响性能的情况下降低基于Transformer的模型的能源消耗和碳排放的效率,来满足对可持续人工智能实践的迫切需求。具体而言,我们使用亚马逊极性数据集进行情感分析,对基于Transformer的模型(BERT、DistilBERT、ALBERT和ELECTRA)应用了剪枝、知识蒸馏和量化。我们还将这些压缩模型的能源效率与本质上碳效率高的Transformer模型(如TinyBERT和MobileBERT)进行了比较。为了评估每个模型的能源消耗和碳排放,我们使用了开源工具CodeCarbon。我们的研究结果表明,应用模型压缩技术后,使用剪枝和蒸馏的BERT能源消耗降低了32.097%,使用剪枝的DistilBERT降低了[公式:见原文]%,使用量化的ALBERT降低了7.12%,使用剪枝和蒸馏的ELECTRA降低了23.934%,同时除了使用量化的ALBERT外,性能指标(准确率、精确率、召回率、F1分数和ROC AUC)保持在95.871 - 99.062%的范围内。具体来说,使用剪枝和蒸馏的BERT准确率达到95.90%,精确率达到95.90%,召回率达到95.90%,F1分数达到95.90%,ROC AUC达到98.87%;使用剪枝的DistilBERT准确率达到95.87%,精确率达到95.87%,召回率达到95.87%,F1分数达到95.87%,ROC AUC达到99.06%;使用剪枝和蒸馏的ELECTRA准确率达到95.92%,精确率达到95.92%,召回率达到95.92%,F1分数达到95.92%,ROC AUC达到99.30%;使用量化的ALBERT准确率达到65.44%,精确率达到67.82%,召回率达到65.44%,F1分数达到63.46%,ROC AUC达到72.31%,这表明由于其已经压缩的架构中的量化敏感性,性能显著下降。总体而言,这证明了使用模型压缩实现可持续人工智能实践的潜力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/8a1e1cf9c4c1/41598_2025_7821_Fig30_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/4f3f8f8992ed/41598_2025_7821_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/3a7f58c67449/41598_2025_7821_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/451253fbd91e/41598_2025_7821_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/3b0610eab040/41598_2025_7821_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/c0b034435be4/41598_2025_7821_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/351828fdd71d/41598_2025_7821_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/0e72f63ef8e1/41598_2025_7821_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/711f5738e44b/41598_2025_7821_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/fb80b3fa6b3d/41598_2025_7821_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/6136859d2017/41598_2025_7821_Figa_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/4f7235709a99/41598_2025_7821_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/09b114b27d5d/41598_2025_7821_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/a10a8b393aea/41598_2025_7821_Fig12_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/457e89565366/41598_2025_7821_Fig13_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/0ba6728483db/41598_2025_7821_Fig14_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/8d0f116e2f75/41598_2025_7821_Fig15_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/74f101d33cd3/41598_2025_7821_Fig16_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/23d5b786f5c3/41598_2025_7821_Fig17_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/b5659c0da80f/41598_2025_7821_Fig18_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/90a55a5e1037/41598_2025_7821_Fig19_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/28d592b484f7/41598_2025_7821_Fig20_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/5c023cd88762/41598_2025_7821_Fig21_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/ccd6898c027b/41598_2025_7821_Fig22_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/1f70c10268ae/41598_2025_7821_Fig23_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/a12cc748be60/41598_2025_7821_Fig24_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/ca5067fb634b/41598_2025_7821_Fig25_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/cd191f4bf13c/41598_2025_7821_Fig26_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/63e6205b5807/41598_2025_7821_Fig27_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/1324a7317990/41598_2025_7821_Fig28_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/5cc43788f8c5/41598_2025_7821_Fig29_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/8a1e1cf9c4c1/41598_2025_7821_Fig30_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/4f3f8f8992ed/41598_2025_7821_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/3a7f58c67449/41598_2025_7821_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/451253fbd91e/41598_2025_7821_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/3b0610eab040/41598_2025_7821_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/c0b034435be4/41598_2025_7821_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/351828fdd71d/41598_2025_7821_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/0e72f63ef8e1/41598_2025_7821_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/711f5738e44b/41598_2025_7821_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/fb80b3fa6b3d/41598_2025_7821_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/6136859d2017/41598_2025_7821_Figa_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/4f7235709a99/41598_2025_7821_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/09b114b27d5d/41598_2025_7821_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/a10a8b393aea/41598_2025_7821_Fig12_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/457e89565366/41598_2025_7821_Fig13_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/0ba6728483db/41598_2025_7821_Fig14_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/8d0f116e2f75/41598_2025_7821_Fig15_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/74f101d33cd3/41598_2025_7821_Fig16_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/23d5b786f5c3/41598_2025_7821_Fig17_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/b5659c0da80f/41598_2025_7821_Fig18_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/90a55a5e1037/41598_2025_7821_Fig19_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/28d592b484f7/41598_2025_7821_Fig20_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/5c023cd88762/41598_2025_7821_Fig21_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/ccd6898c027b/41598_2025_7821_Fig22_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/1f70c10268ae/41598_2025_7821_Fig23_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/a12cc748be60/41598_2025_7821_Fig24_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/ca5067fb634b/41598_2025_7821_Fig25_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/cd191f4bf13c/41598_2025_7821_Fig26_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/63e6205b5807/41598_2025_7821_Fig27_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/1324a7317990/41598_2025_7821_Fig28_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/5cc43788f8c5/41598_2025_7821_Fig29_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4be5/12222641/8a1e1cf9c4c1/41598_2025_7821_Fig30_HTML.jpg

相似文献

1
Comparative analysis of model compression techniques for achieving carbon efficient AI.实现碳高效人工智能的模型压缩技术比较分析
Sci Rep. 2025 Jul 2;15(1):23461. doi: 10.1038/s41598-025-07821-w.
2
Supervised Machine Learning Models for Predicting Sepsis-Associated Liver Injury in Patients With Sepsis: Development and Validation Study Based on a Multicenter Cohort Study.用于预测脓毒症患者脓毒症相关肝损伤的监督式机器学习模型:基于多中心队列研究的开发与验证研究
J Med Internet Res. 2025 May 26;27:e66733. doi: 10.2196/66733.
3
Comparative analysis of convolutional neural networks and transformer architectures for breast cancer histopathological image classification.用于乳腺癌组织病理学图像分类的卷积神经网络与Transformer架构的比较分析
Front Med (Lausanne). 2025 Jun 17;12:1606336. doi: 10.3389/fmed.2025.1606336. eCollection 2025.
4
Are Current Survival Prediction Tools Useful When Treating Subsequent Skeletal-related Events From Bone Metastases?当前的生存预测工具在治疗骨转移后的骨骼相关事件时有用吗?
Clin Orthop Relat Res. 2024 Sep 1;482(9):1710-1721. doi: 10.1097/CORR.0000000000003030. Epub 2024 Mar 22.
5
Knowledge Graph-Enhanced Deep Learning Model (H-SYSTEM) for Hypertensive Intracerebral Hemorrhage: Model Development and Validation.用于高血压性脑出血的知识图谱增强深度学习模型(H-SYSTEM):模型开发与验证
J Med Internet Res. 2025 Jun 12;27:e66055. doi: 10.2196/66055.
6
Optimising TinyML with quantization and distillation of transformer and mamba models for indoor localisation on edge devices.通过对变压器和曼巴模型进行量化和蒸馏来优化用于边缘设备室内定位的 TinyML。
Sci Rep. 2025 Mar 24;15(1):10081. doi: 10.1038/s41598-025-94205-9.
7
A deep learning approach to direct immunofluorescence pattern recognition in autoimmune bullous diseases.深度学习方法在自身免疫性大疱性疾病中的直接免疫荧光模式识别。
Br J Dermatol. 2024 Jul 16;191(2):261-266. doi: 10.1093/bjd/ljae142.
8
Signs and symptoms to determine if a patient presenting in primary care or hospital outpatient settings has COVID-19.在基层医疗机构或医院门诊环境中,如果患者出现以下症状和体征,可判断其是否患有 COVID-19。
Cochrane Database Syst Rev. 2022 May 20;5(5):CD013665. doi: 10.1002/14651858.CD013665.pub3.
9
Harnessing Moderate-Sized Language Models for Reliable Patient Data Deidentification in Emergency Department Records: Algorithm Development, Validation, and Implementation Study.利用中等规模语言模型对急诊科记录中的患者数据进行可靠去识别:算法开发、验证与实施研究。
JMIR AI. 2025 Apr 1;4:e57828. doi: 10.2196/57828.
10
Algorithmic Classification of Psychiatric Disorder-Related Spontaneous Communication Using Large Language Model Embeddings: Algorithm Development and Validation.使用大语言模型嵌入对精神障碍相关自发交流进行算法分类:算法开发与验证
JMIR AI. 2025 May 30;4:e67369. doi: 10.2196/67369.

本文引用的文献

1
A Survey on Deep Neural Network Pruning: Taxonomy, Comparison, Analysis, and Recommendations.深度神经网络剪枝研究综述:分类、比较、分析与建议
IEEE Trans Pattern Anal Mach Intell. 2024 Dec;46(12):10558-10578. doi: 10.1109/TPAMI.2024.3447085. Epub 2024 Nov 6.
2
Pruning and quantization algorithm with applications in memristor-based convolutional neural network.基于忆阻器的卷积神经网络中的剪枝与量化算法及其应用
Cogn Neurodyn. 2024 Feb;18(1):233-245. doi: 10.1007/s11571-022-09927-7. Epub 2023 Jan 19.
3
Mitigating carbon footprint for knowledge distillation based deep learning model compression.
减轻基于知识蒸馏的深度学习模型压缩的碳足迹。
PLoS One. 2023 May 15;18(5):e0285668. doi: 10.1371/journal.pone.0285668. eCollection 2023.
4
Automatic Whole Slide Pathology Image Diagnosis Framework via Unit Stochastic Selection and Attention Fusion.基于单元随机选择和注意力融合的全玻片病理图像自动诊断框架
Neurocomputing (Amst). 2021 Sep 17;453:312-325. doi: 10.1016/j.neucom.2020.04.153. Epub 2021 Jan 23.
5
Non-Structured DNN Weight Pruning-Is It Beneficial in Any Platform?非结构化深度神经网络权重剪枝——在任何平台上都有益吗?
IEEE Trans Neural Netw Learn Syst. 2022 Sep;33(9):4930-4944. doi: 10.1109/TNNLS.2021.3063265. Epub 2022 Aug 31.