• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

中等规模的蛋白质语言模型在真实数据集上的迁移学习中表现良好。

Medium-sized protein language models perform well at transfer learning on realistic datasets.

作者信息

Vieira Luiz C, Handojo Morgan L, Wilke Claus O

机构信息

Department of Integrative Biology, The University of Texas at Austin, Austin, TX, USA.

出版信息

Sci Rep. 2025 Jul 1;15(1):21400. doi: 10.1038/s41598-025-05674-x.

DOI:10.1038/s41598-025-05674-x
PMID:40594749
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12217344/
Abstract

Protein language models (pLMs) can offer deep insights into evolutionary and structural properties of proteins. While larger models, such as the 15 billion parameter model ESM-2, promise to capture more complex patterns in sequence space, they also present practical challenges due to their high dimensionality and high computational cost. We systematically evaluated the performance of various ESM-style models across multiple biological datasets to assess the impact of model size on transfer learning via feature extraction. Surprisingly, we found that larger models do not necessarily outperform smaller ones, in particular when data is limited. Medium-sized models, such as ESM-2 650M and ESM C 600M, demonstrated consistently good performance, falling only slightly behind their larger counterparts-ESM-2 15B and ESM C 6B-despite being many times smaller. Additionally, we compared various methods of compressing embeddings prior to transfer learning, and we found that mean embeddings consistently outperformed other compression methods. In summary, ESM C 600M with mean embeddings offers an optimal balance between performance and efficiency, making it a practical and scalable choice for transfer learning in realistic biological applications.

摘要

蛋白质语言模型(pLMs)能够深入洞察蛋白质的进化和结构特性。虽然更大的模型,如拥有150亿参数的模型ESM-2,有望捕捉序列空间中更复杂的模式,但由于其高维度和高计算成本,也带来了实际挑战。我们系统地评估了各种ESM风格模型在多个生物学数据集上的性能,以评估模型大小对通过特征提取进行迁移学习的影响。令人惊讶的是,我们发现更大的模型不一定比小模型表现更好,特别是在数据有限的情况下。中等大小的模型,如ESM-2 650M和ESM C 600M,表现出始终如一的良好性能,尽管比它们更大的对应模型ESM-2 15B和ESM C 6B小很多倍,但仅略落后于它们。此外,我们比较了迁移学习前压缩嵌入的各种方法,发现平均嵌入始终优于其他压缩方法。总之,具有平均嵌入的ESM C 600M在性能和效率之间提供了最佳平衡,使其成为实际生物学应用中迁移学习的实用且可扩展的选择。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b2a/12217344/c22bb4468685/41598_2025_5674_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b2a/12217344/171b8526585c/41598_2025_5674_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b2a/12217344/f13317ffc5fe/41598_2025_5674_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b2a/12217344/b85d768c9fed/41598_2025_5674_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b2a/12217344/6053d85b1e96/41598_2025_5674_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b2a/12217344/c22bb4468685/41598_2025_5674_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b2a/12217344/171b8526585c/41598_2025_5674_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b2a/12217344/f13317ffc5fe/41598_2025_5674_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b2a/12217344/b85d768c9fed/41598_2025_5674_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b2a/12217344/6053d85b1e96/41598_2025_5674_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8b2a/12217344/c22bb4468685/41598_2025_5674_Fig5_HTML.jpg

相似文献

1
Medium-sized protein language models perform well at transfer learning on realistic datasets.中等规模的蛋白质语言模型在真实数据集上的迁移学习中表现良好。
Sci Rep. 2025 Jul 1;15(1):21400. doi: 10.1038/s41598-025-05674-x.
2
Scaling down for efficiency: Medium-sized protein language models perform well at transfer learning on realistic datasets.为提高效率而缩小规模:中型蛋白质语言模型在真实数据集的迁移学习中表现良好。
bioRxiv. 2025 Jan 28:2024.11.22.624936. doi: 10.1101/2024.11.22.624936.
3
Comparison of Two Modern Survival Prediction Tools, SORG-MLA and METSSS, in Patients With Symptomatic Long-bone Metastases Who Underwent Local Treatment With Surgery Followed by Radiotherapy and With Radiotherapy Alone.两种现代生存预测工具 SORG-MLA 和 METSSS 在接受手术联合放疗和单纯放疗治疗有症状长骨转移患者中的比较。
Clin Orthop Relat Res. 2024 Dec 1;482(12):2193-2208. doi: 10.1097/CORR.0000000000003185. Epub 2024 Jul 23.
4
Leveraging a foundation model zoo for cell similarity search in oncological microscopy across devices.利用基础模型库进行跨设备肿瘤显微镜检查中的细胞相似性搜索。
Front Oncol. 2025 Jun 18;15:1480384. doi: 10.3389/fonc.2025.1480384. eCollection 2025.
5
Is It Possible to Develop a Patient-reported Experience Measure With Lower Ceiling Effect?是否有可能开发一种天花板效应较低的患者报告体验测量方法?
Clin Orthop Relat Res. 2025 Apr 1;483(4):693-703. doi: 10.1097/CORR.0000000000003262. Epub 2024 Oct 25.
6
Systemic pharmacological treatments for chronic plaque psoriasis: a network meta-analysis.系统性药理学治疗慢性斑块状银屑病:网络荟萃分析。
Cochrane Database Syst Rev. 2021 Apr 19;4(4):CD011535. doi: 10.1002/14651858.CD011535.pub4.
7
Systemic pharmacological treatments for chronic plaque psoriasis: a network meta-analysis.慢性斑块状银屑病的全身药理学治疗:一项网状Meta分析。
Cochrane Database Syst Rev. 2020 Jan 9;1(1):CD011535. doi: 10.1002/14651858.CD011535.pub3.
8
Systemic pharmacological treatments for chronic plaque psoriasis: a network meta-analysis.慢性斑块状银屑病的全身药理学治疗:一项网状荟萃分析。
Cochrane Database Syst Rev. 2017 Dec 22;12(12):CD011535. doi: 10.1002/14651858.CD011535.pub2.
9
MoRF_ESM: Prediction of MoRFs in disordered proteins based on a deep transformer protein language model.MoRF_ESM:基于深度变压器蛋白质语言模型预测无序蛋白质中的分子识别特征片段
J Bioinform Comput Biol. 2024 Apr;22(2):2450006. doi: 10.1142/S0219720024500069. Epub 2024 May 28.
10
Signs and symptoms to determine if a patient presenting in primary care or hospital outpatient settings has COVID-19.在基层医疗机构或医院门诊环境中,如果患者出现以下症状和体征,可判断其是否患有 COVID-19。
Cochrane Database Syst Rev. 2022 May 20;5(5):CD013665. doi: 10.1002/14651858.CD013665.pub3.

引用本文的文献

1
Mechanistic modeling or machine learning for detecting variants of concern: Why not both?用于检测关注变体的机制建模或机器学习:为何不两者兼用?
Proc Natl Acad Sci U S A. 2025 Jul 29;122(30):e2513608122. doi: 10.1073/pnas.2513608122. Epub 2025 Jul 21.

本文引用的文献

1
A systematic evaluation of the language-of-viral-escape model using multiple machine learning frameworks.使用多个机器学习框架对病毒逃逸模型语言进行的系统评估。
J R Soc Interface. 2025 Apr;22(225):20240598. doi: 10.1098/rsif.2024.0598. Epub 2025 Apr 30.
2
Aggregating residue-level protein language model embeddings with optimal transport.通过最优传输聚合残基水平的蛋白质语言模型嵌入
Bioinform Adv. 2025 Mar 20;5(1):vbaf060. doi: 10.1093/bioadv/vbaf060. eCollection 2025.
3
Simulating 500 million years of evolution with a language model.
用语言模型模拟5亿年的进化历程。
Science. 2025 Feb 21;387(6736):850-858. doi: 10.1126/science.ads0018. Epub 2025 Jan 16.
4
Semantic search using protein large language models detects class II microcins in bacterial genomes.基于蛋白质大型语言模型的语义搜索可在细菌基因组中检测到 II 类微菌素。
mSystems. 2024 Oct 22;9(10):e0104424. doi: 10.1128/msystems.01044-24. Epub 2024 Sep 18.
5
Deep mutational scanning and machine learning for the analysis of antimicrobial-peptide features driving membrane selectivity.利用深度突变扫描和机器学习分析驱动膜选择性的抗菌肽特征。
Nat Biomed Eng. 2024 Jul;8(7):842-853. doi: 10.1038/s41551-024-01243-1. Epub 2024 Jul 31.
6
Democratizing protein language models with parameter-efficient fine-tuning.参数高效微调:用民主化方法对蛋白质语言模型进行优化。
Proc Natl Acad Sci U S A. 2024 Jun 25;121(26):e2405840121. doi: 10.1073/pnas.2405840121. Epub 2024 Jun 20.
7
Correcting PCR amplification errors in unique molecular identifiers to generate accurate numbers of sequencing molecules.校正独特分子标识符中的PCR扩增错误以生成准确的测序分子数量。
Nat Methods. 2024 Mar;21(3):401-405. doi: 10.1038/s41592-024-02168-y. Epub 2024 Feb 5.
8
ProGen2: Exploring the boundaries of protein language models.ProGen2:探索蛋白质语言模型的边界。
Cell Syst. 2023 Nov 15;14(11):968-978.e3. doi: 10.1016/j.cels.2023.10.002. Epub 2023 Oct 30.
9
De novo design of protein structure and function with RFdiffusion.利用 RFdiffusion 从头设计蛋白质结构和功能。
Nature. 2023 Aug;620(7976):1089-1100. doi: 10.1038/s41586-023-06415-8. Epub 2023 Jul 11.
10
Evolutionary-scale prediction of atomic-level protein structure with a language model.用语言模型进行原子级蛋白质结构的进化尺度预测。
Science. 2023 Mar 17;379(6637):1123-1130. doi: 10.1126/science.ade2574. Epub 2023 Mar 16.