• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

Nach0:多模态自然与化学语言基础模型。

nach0: multimodal natural and chemical languages foundation model.

作者信息

Livne Micha, Miftahutdinov Zulfat, Tutubalina Elena, Kuznetsov Maksim, Polykovskiy Daniil, Brundyn Annika, Jhunjhunwala Aastha, Costa Anthony, Aliper Alex, Aspuru-Guzik Alán, Zhavoronkov Alex

机构信息

NVIDIA 2788 San Tomas Expressway Santa Clara 95051 CA USA.

Insilico Medicine Canada Inc. 3710-1250 René-Lévesque West Montreal Quebec Canada.

出版信息

Chem Sci. 2024 May 8;15(22):8380-8389. doi: 10.1039/d4sc00966e. eCollection 2024 Jun 5.

DOI:10.1039/d4sc00966e
PMID:38846388
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11151847/
Abstract

Large Language Models (LLMs) have substantially driven scientific progress in various domains, and many papers have demonstrated their ability to tackle complex problems with creative solutions. Our paper introduces a new foundation model, nach0, capable of solving various chemical and biological tasks: biomedical question answering, named entity recognition, molecular generation, molecular synthesis, attributes prediction, and others. nach0 is a multi-domain and multi-task encoder-decoder LLM pre-trained on unlabeled text from scientific literature, patents, and molecule strings to incorporate a range of chemical and linguistic knowledge. We employed instruction tuning, where specific task-related instructions are utilized to fine-tune nach0 for the final set of tasks. To train nach0 effectively, we leverage the NeMo framework, enabling efficient parallel optimization of both base and large model versions. Extensive experiments demonstrate that our model outperforms state-of-the-art baselines on single-domain and cross-domain tasks. Furthermore, it can generate high-quality outputs in molecular and textual formats, showcasing its effectiveness in multi-domain setups.

摘要

大语言模型(LLMs)在各个领域极大地推动了科学进步,许多论文都展示了它们用创造性解决方案解决复杂问题的能力。我们的论文介绍了一种新的基础模型nach0,它能够解决各种化学和生物学任务:生物医学问答、命名实体识别、分子生成、分子合成、属性预测等。nach0是一个多领域、多任务的编码器-解码器大语言模型,它在来自科学文献、专利和分子字符串的未标记文本上进行预训练,以融入一系列化学和语言知识。我们采用了指令微调,即利用特定的与任务相关的指令对nach0进行微调,以完成最终的任务集。为了有效地训练nach0,我们利用了NeMo框架,实现了基础模型和大型模型版本的高效并行优化。大量实验表明,我们的模型在单领域和跨领域任务上优于当前的基准模型。此外,它能够生成高质量的分子和文本格式的输出,展示了其在多领域设置中的有效性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/97d7/11151847/30c47666c1c4/d4sc00966e-f4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/97d7/11151847/95e508e8b273/d4sc00966e-f1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/97d7/11151847/3a3eebe543f4/d4sc00966e-f2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/97d7/11151847/ff6287517107/d4sc00966e-f3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/97d7/11151847/30c47666c1c4/d4sc00966e-f4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/97d7/11151847/95e508e8b273/d4sc00966e-f1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/97d7/11151847/3a3eebe543f4/d4sc00966e-f2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/97d7/11151847/ff6287517107/d4sc00966e-f3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/97d7/11151847/30c47666c1c4/d4sc00966e-f4.jpg

相似文献

1
nach0: multimodal natural and chemical languages foundation model.Nach0:多模态自然与化学语言基础模型。
Chem Sci. 2024 May 8;15(22):8380-8389. doi: 10.1039/d4sc00966e. eCollection 2024 Jun 5.
2
Taiyi: a bilingual fine-tuned large language model for diverse biomedical tasks.太乙:一个用于多种生物医学任务的双语精调大型语言模型。
J Am Med Inform Assoc. 2024 Sep 1;31(9):1865-1874. doi: 10.1093/jamia/ocae037.
3
BioInstruct: instruction tuning of large language models for biomedical natural language processing.BioInstruct:用于生物医学自然语言处理的大型语言模型的指令调整。
J Am Med Inform Assoc. 2024 Sep 1;31(9):1821-1832. doi: 10.1093/jamia/ocae122.
4
Benchmarking for biomedical natural language processing tasks with a domain specific ALBERT.基于领域特定的 ALBERT 进行生物医学自然语言处理任务的基准测试。
BMC Bioinformatics. 2022 Apr 21;23(1):144. doi: 10.1186/s12859-022-04688-w.
5
A comprehensive evaluation of large Language models on benchmark biomedical text processing tasks.对基准生物医学文本处理任务中大型语言模型的全面评估。
Comput Biol Med. 2024 Mar;171:108189. doi: 10.1016/j.compbiomed.2024.108189. Epub 2024 Feb 20.
6
LEAP: LLM instruction-example adaptive prompting framework for biomedical relation extraction.LEAP:用于生物医学关系抽取的 LLM 指令-示例自适应提示框架。
J Am Med Inform Assoc. 2024 Sep 1;31(9):2010-2018. doi: 10.1093/jamia/ocae147.
7
BioBERT: a pre-trained biomedical language representation model for biomedical text mining.BioBERT:一种用于生物医学文本挖掘的预训练生物医学语言表示模型。
Bioinformatics. 2020 Feb 15;36(4):1234-1240. doi: 10.1093/bioinformatics/btz682.
8
MedChatZH: A tuning LLM for traditional Chinese medicine consultations.医聊 ChatZH:一个用于中医咨询的调优大语言模型。
Comput Biol Med. 2024 Apr;172:108290. doi: 10.1016/j.compbiomed.2024.108290. Epub 2024 Mar 13.
9
Advancing Multimodal Large Language Models in Chart Question Answering with Visualization-Referenced Instruction Tuning.通过可视化参考指令微调推进图表问答中的多模态大语言模型
IEEE Trans Vis Comput Graph. 2025 Jan;31(1):525-535. doi: 10.1109/TVCG.2024.3456159. Epub 2024 Nov 25.
10
Few-shot biomedical named entity recognition via knowledge-guided instance generation and prompt contrastive learning.通过知识引导实例生成和提示对比学习的少样本生物医学命名实体识别。
Bioinformatics. 2023 Aug 1;39(8). doi: 10.1093/bioinformatics/btad496.

引用本文的文献

1
Chemical Language Model Linker: Blending Text and Molecules with Modular Adapters.化学语言模型链接器:通过模块化适配器融合文本与分子
J Chem Inf Model. 2025 Sep 8;65(17):8944-8956. doi: 10.1021/acs.jcim.5c00853. Epub 2025 Aug 21.
2
AI and automation: democratizing automation and the evolution towards true AI-autonomous robotics.人工智能与自动化:使自动化民主化以及向真正的人工智能自主机器人技术的演进。
Chem Sci. 2025 Aug 4. doi: 10.1039/d5sc03183d.
3
RPSubAlign: a novel sequence-based molecular representation method for retrosynthesis prediction with improved validity and robustness.

本文引用的文献

1
An extensive benchmark study on biomedical text generation and mining with ChatGPT.一项关于使用ChatGPT进行生物医学文本生成和挖掘的广泛基准研究。
Bioinformatics. 2023 Sep 2;39(9). doi: 10.1093/bioinformatics/btad557.
2
Prediction of Clinical Trials Outcomes Based on Target Choice and Clinical Trial Design with Multi-Modal Artificial Intelligence.基于多模态人工智能的靶点选择和临床试验设计对临床试验结果的预测
Clin Pharmacol Ther. 2023 Nov;114(5):972-980. doi: 10.1002/cpt.3008. Epub 2023 Aug 10.
3
Chemistry42: An AI-Driven Platform for Molecular Design and Optimization.
RPSubAlign:一种基于序列的新型分子表示方法,用于逆合成预测,具有更高的有效性和鲁棒性。
Brief Bioinform. 2025 May 1;26(3). doi: 10.1093/bib/bbaf257.
4
Chemical Language Model Linker: blending text and molecules with modular adapters.化学语言模型链接器:通过模块化适配器融合文本与分子。
ArXiv. 2025 Jun 13:arXiv:2410.20182v3.
5
A Perspective on Foundation Models in Chemistry.化学领域基础模型的视角
JACS Au. 2025 Mar 25;5(4):1499-1518. doi: 10.1021/jacsau.4c01160. eCollection 2025 Apr 28.
6
Cross-disciplinary perspectives on the potential for artificial intelligence across chemistry.关于人工智能在化学领域潜力的跨学科观点。
Chem Soc Rev. 2025 Apr 25. doi: 10.1039/d5cs00146c.
7
Unified Deep Learning of Molecular and Protein Language Representations with T5ProtChem.利用T5ProtChem对分子和蛋白质语言表征进行统一深度学习。
J Chem Inf Model. 2025 Apr 28;65(8):3990-3998. doi: 10.1021/acs.jcim.5c00051. Epub 2025 Apr 8.
8
Revolutionizing oncology: the role of Artificial Intelligence (AI) as an antibody design, and optimization tools.肿瘤学的变革:人工智能(AI)作为抗体设计与优化工具的作用。
Biomark Res. 2025 Mar 29;13(1):52. doi: 10.1186/s40364-025-00764-4.
9
A review of large language models and autonomous agents in chemistry.化学领域中大型语言模型与自主智能体的综述。
Chem Sci. 2024 Dec 9;16(6):2514-2572. doi: 10.1039/d4sc03921a. eCollection 2025 Feb 5.
10
An automatic end-to-end chemical synthesis development platform powered by large language models.基于大型语言模型的自动化端到端化学合成开发平台。
Nat Commun. 2024 Nov 23;15(1):10160. doi: 10.1038/s41467-024-54457-x.
Chemistry42:一个人工智能驱动的分子设计和优化平台。
J Chem Inf Model. 2023 Feb 13;63(3):695-701. doi: 10.1021/acs.jcim.2c01191. Epub 2023 Feb 2.
4
SELFIES and the future of molecular string representations.自拍与分子串表示法的未来。
Patterns (N Y). 2022 Oct 14;3(10):100588. doi: 10.1016/j.patter.2022.100588.
5
BioGPT: generative pre-trained transformer for biomedical text generation and mining.BioGPT:用于生物医学文本生成和挖掘的生成式预训练转换器。
Brief Bioinform. 2022 Nov 19;23(6). doi: 10.1093/bib/bbac409.
6
Language models can learn complex molecular distributions.语言模型可以学习复杂的分子分布。
Nat Commun. 2022 Jun 7;13(1):3293. doi: 10.1038/s41467-022-30839-x.
7
Medical concept normalization in clinical trials with drug and disease representation learning.临床试验中基于药物和疾病表示学习的医学概念规范化。
Bioinformatics. 2021 Nov 5;37(21):3856-3864. doi: 10.1093/bioinformatics/btab474.
8
Molecular Sets (MOSES): A Benchmarking Platform for Molecular Generation Models.分子集(MOSES):分子生成模型的基准测试平台。
Front Pharmacol. 2020 Dec 18;11:565644. doi: 10.3389/fphar.2020.565644. eCollection 2020.
9
Molecular Generation for Desired Transcriptome Changes With Adversarial Autoencoders.使用对抗自编码器实现所需转录组变化的分子生成
Front Pharmacol. 2020 Apr 17;11:269. doi: 10.3389/fphar.2020.00269. eCollection 2020.
10
BioBERT: a pre-trained biomedical language representation model for biomedical text mining.BioBERT:一种用于生物医学文本挖掘的预训练生物医学语言表示模型。
Bioinformatics. 2020 Feb 15;36(4):1234-1240. doi: 10.1093/bioinformatics/btz682.