• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于生成式检索增强本体图和多智能体策略的基于解释性大语言模型的材料设计

Generative Retrieval-Augmented Ontologic Graph and Multiagent Strategies for Interpretive Large Language Model-Based Materials Design.

作者信息

Buehler Markus J

机构信息

Laboratory for Atomistic and Molecular Mechanics (LAMM), Massachusetts Institute of Technology, 77 Massachusetts Ave., Cambridge, Massachusetts 02139, United States.

Department of Civil and Environmental Engineering, Massachusetts Institute of Technology, 77 Massachusetts Ave., Cambridge, Massachusetts 02139, United States.

出版信息

ACS Eng Au. 2024 Jan 12;4(2):241-277. doi: 10.1021/acsengineeringau.3c00058. eCollection 2024 Apr 17.

DOI:10.1021/acsengineeringau.3c00058
PMID:38646516
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11027160/
Abstract

Transformer neural networks show promising capabilities, in particular for uses in materials analysis, design, and manufacturing, including their capacity to work effectively with human language, symbols, code, and numerical data. Here, we explore the use of large language models (LLMs) as a tool that can support engineering analysis of materials, applied to retrieving key information about subject areas, developing research hypotheses, discovery of mechanistic relationships across disparate areas of knowledge, and writing and executing simulation codes for active knowledge generation based on physical ground truths. Moreover, when used as sets of AI agents with specific features, capabilities, and instructions, LLMs can provide powerful problem-solution strategies for applications in analysis and design problems. Our experiments focus on using a fine-tuned model, MechGPT, developed based on training data in the mechanics of materials domain. We first affirm how fine-tuning endows LLMs with a reasonable understanding of subject area knowledge. However, when queried outside the context of learned matter, LLMs can have difficulty recalling correct information and may hallucinate. We show how this can be addressed using retrieval-augmented Ontological Knowledge Graph strategies. The graph-based strategy helps us not only to discern how the model understands what concepts are important but also how they are related, which significantly improves generative performance and also naturally allows for injection of new and augmented data sources into generative AI algorithms. We find that the additional feature of relatedness provides advantages over regular retrieval augmentation approaches and not only improves LLM performance but also provides mechanistic insights for exploration of a material design process. Illustrated for a use case of relating distinct areas of knowledge, here, music and proteins, such strategies can also provide an interpretable graph structure with rich information at the node, edge, and subgraph level that provides specific insights into mechanisms and relationships. We discuss other approaches to improve generative qualities, including nonlinear sampling strategies and agent-based modeling that offer enhancements over single-shot generations, whereby LLMs are used to both generate content and assess content against an objective target. Examples provided include complex question answering, code generation, and execution in the context of automated force-field development from actively learned density functional theory (DFT) modeling and data analysis.

摘要

Transformer神经网络展现出了令人期待的能力,特别是在材料分析、设计和制造中的应用,包括其有效处理人类语言、符号、代码和数值数据的能力。在此,我们探索将大语言模型(LLMs)用作一种工具,以支持材料的工程分析,应用于检索有关主题领域的关键信息、提出研究假设、发现不同知识领域之间的机理关系,以及编写和执行基于物理基本事实的用于主动知识生成的模拟代码。此外,当用作具有特定特征、能力和指令的人工智能代理集时,大语言模型可以为分析和设计问题的应用提供强大的问题解决策略。我们的实验聚焦于使用基于材料力学领域训练数据开发的微调模型MechGPT。我们首先确认微调如何赋予大语言模型对主题领域知识的合理理解。然而,当在所学内容的上下文之外进行查询时,大语言模型可能难以回忆起正确信息并可能产生幻觉。我们展示了如何使用检索增强的本体知识图谱策略来解决这一问题。基于图谱的策略不仅帮助我们辨别模型如何理解哪些概念是重要的,还能了解它们之间的关系,这显著提高了生成性能,并且自然地允许将新的和增强的数据源注入到生成式人工智能算法中。我们发现相关性这一附加特征比常规检索增强方法具有优势,不仅提高了大语言模型的性能,还为材料设计过程的探索提供了机理见解。在此以关联音乐和蛋白质等不同知识领域的用例进行说明,此类策略还可以提供一个在节点、边和子图层面具有丰富信息的可解释图谱结构,从而对机理和关系提供具体见解。我们讨论了其他提高生成质量的方法,包括非线性采样策略和基于代理的建模,这些方法比单次生成提供了改进,即大语言模型用于生成内容并根据客观目标评估内容。提供的示例包括在从主动学习的密度泛函理论(DFT)建模和数据分析进行自动力场开发的背景下的复杂问答、代码生成和执行。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/7c335607d941/eg3c00058_0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/7537aa37addc/eg3c00058_0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/c3d2cef2c0c4/eg3c00058_0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/924585eafd4a/eg3c00058_0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/f3bba7ae2322/eg3c00058_0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/e732a2892f07/eg3c00058_0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/35db2644b44f/eg3c00058_0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/b3b11c06ef3f/eg3c00058_0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/7c335607d941/eg3c00058_0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/7537aa37addc/eg3c00058_0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/c3d2cef2c0c4/eg3c00058_0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/924585eafd4a/eg3c00058_0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/f3bba7ae2322/eg3c00058_0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/e732a2892f07/eg3c00058_0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/35db2644b44f/eg3c00058_0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/b3b11c06ef3f/eg3c00058_0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d926/11027160/7c335607d941/eg3c00058_0008.jpg

相似文献

1
Generative Retrieval-Augmented Ontologic Graph and Multiagent Strategies for Interpretive Large Language Model-Based Materials Design.基于生成式检索增强本体图和多智能体策略的基于解释性大语言模型的材料设计
ACS Eng Au. 2024 Jan 12;4(2):241-277. doi: 10.1021/acsengineeringau.3c00058. eCollection 2024 Apr 17.
2
Large Language Models Can Enable Inductive Thematic Analysis of a Social Media Corpus in a Single Prompt: Human Validation Study.大语言模型可通过单一提示实现社交媒体语料库的归纳主题分析:人类验证研究。
JMIR Infodemiology. 2024 Aug 29;4:e59641. doi: 10.2196/59641.
3
Learning to Make Rare and Complex Diagnoses With Generative AI Assistance: Qualitative Study of Popular Large Language Models.利用生成式人工智能辅助学习罕见且复杂的诊断:对流行的大型语言模型的定性研究。
JMIR Med Educ. 2024 Feb 13;10:e51391. doi: 10.2196/51391.
4
Optimizing large language models in digestive disease: strategies and challenges to improve clinical outcomes.优化消化疾病中的大语言模型:改善临床结局的策略和挑战。
Liver Int. 2024 Sep;44(9):2114-2124. doi: 10.1111/liv.15974. Epub 2024 May 31.
5
Taiyi: a bilingual fine-tuned large language model for diverse biomedical tasks.太乙:一个用于多种生物医学任务的双语精调大型语言模型。
J Am Med Inform Assoc. 2024 Sep 1;31(9):1865-1874. doi: 10.1093/jamia/ocae037.
6
An Empirical Evaluation of Prompting Strategies for Large Language Models in Zero-Shot Clinical Natural Language Processing: Algorithm Development and Validation Study.零样本临床自然语言处理中大型语言模型提示策略的实证评估:算法开发与验证研究
JMIR Med Inform. 2024 Apr 8;12:e55318. doi: 10.2196/55318.
7
MedExpQA: Multilingual benchmarking of Large Language Models for Medical Question Answering.MedExpQA:用于医学问答的大型语言模型的多语言基准测试。
Artif Intell Med. 2024 Sep;155:102938. doi: 10.1016/j.artmed.2024.102938. Epub 2024 Jul 31.
8
Evaluation of the Performance of Generative AI Large Language Models ChatGPT, Google Bard, and Microsoft Bing Chat in Supporting Evidence-Based Dentistry: Comparative Mixed Methods Study.评估生成式 AI 大语言模型 ChatGPT、Google Bard 和 Microsoft Bing Chat 在支持循证牙科方面的性能:比较混合方法研究。
J Med Internet Res. 2023 Dec 28;25:e51580. doi: 10.2196/51580.
9
Almanac - Retrieval-Augmented Language Models for Clinical Medicine.用于临床医学的年鉴检索增强语言模型。
NEJM AI. 2024 Feb;1(2). doi: 10.1056/aioa2300068. Epub 2024 Jan 25.
10
Large language models as tax attorneys: a case study in legal capabilities emergence.大语言模型当税法顾问:法律能力涌现的案例研究。
Philos Trans A Math Phys Eng Sci. 2024 Apr 15;382(2270):20230159. doi: 10.1098/rsta.2023.0159. Epub 2024 Feb 26.

引用本文的文献

1
Graph retrieval augmented large language models for facial phenotype associated rare genetic disease.用于面部表型相关罕见遗传病的图谱检索增强大语言模型
NPJ Digit Med. 2025 Aug 24;8(1):543. doi: 10.1038/s41746-025-01955-x.
2
Molecular analysis and design using generative artificial intelligence multi-agent modeling.使用生成式人工智能多智能体建模的分子分析与设计
Mol Syst Des Eng. 2025 Jan 24;10(4):314-337. doi: 10.1039/d4me00174e. eCollection 2025 Mar 31.
3
Automating alloy design and discovery with physics-aware multimodal multiagent AI.

本文引用的文献

1
BioinspiredLLM: Conversational Large Language Model for the Mechanics of Biological and Bio-Inspired Materials.受生物启发的语言模型:用于生物及受生物启发材料力学的对话式大语言模型
Adv Sci (Weinh). 2024 Mar;11(10):e2306724. doi: 10.1002/advs.202306724. Epub 2023 Dec 25.
2
Microwave synthesis of molybdenene from MoS.由二硫化钼微波合成钼烯
Nat Nanotechnol. 2023 Dec;18(12):1430-1438. doi: 10.1038/s41565-023-01484-2. Epub 2023 Sep 4.
3
Generative design of proteins based on secondary structure constraints using an attention-based diffusion model.
利用物理感知多模态多智能体人工智能实现合金设计与发现的自动化。
Proc Natl Acad Sci U S A. 2025 Jan 28;122(4):e2414074122. doi: 10.1073/pnas.2414074122. Epub 2025 Jan 24.
4
SciAgents: Automating Scientific Discovery Through Bioinspired Multi-Agent Intelligent Graph Reasoning.科学智能体:通过受生物启发的多智能体智能图推理实现科学发现自动化
Adv Mater. 2025 Jun;37(22):e2413523. doi: 10.1002/adma.202413523. Epub 2024 Dec 18.
5
ProtAgents: protein discovery large language model multi-agent collaborations combining physics and machine learning.ProtAgents:蛋白质发现大型语言模型,结合物理和机器学习的多智能体协作。
Digit Discov. 2024 May 17;3(7):1389-1409. doi: 10.1039/d4dd00013g. eCollection 2024 Jul 10.
基于二级结构约束,使用基于注意力的扩散模型进行蛋白质的生成式设计。
Chem. 2023 Jul 13;9(7):1828-1849. doi: 10.1016/j.chempr.2023.03.020. Epub 2023 Apr 20.
4
Modeling and design of heterogeneous hierarchical bioinspired spider web structures using deep learning and additive manufacturing.基于深度学习和增材制造的异构分层仿生蜘蛛结构的建模与设计。
Proc Natl Acad Sci U S A. 2023 Aug;120(31):e2305273120. doi: 10.1073/pnas.2305273120. Epub 2023 Jul 24.
5
Unsupervised cross-domain translation via deep learning and adversarial attention neural networks and application to music-inspired protein designs.通过深度学习和对抗性注意力神经网络进行无监督跨域翻译及其在音乐启发的蛋白质设计中的应用。
Patterns (N Y). 2023 Feb 14;4(3):100692. doi: 10.1016/j.patter.2023.100692. eCollection 2023 Mar 10.
6
Large language models and the perils of their hallucinations.大语言模型及其幻觉的风险。
Crit Care. 2023 Mar 21;27(1):120. doi: 10.1186/s13054-023-04393-x.
7
Attention is not all you need: the complicated case of ethically using large language models in healthcare and medicine.注意力并非全部所需:在医疗保健和医学中使用大型语言模型所涉及的复杂伦理问题。
EBioMedicine. 2023 Apr;90:104512. doi: 10.1016/j.ebiom.2023.104512. Epub 2023 Mar 15.
8
End-to-End Protein Normal Mode Frequency Predictions Using Language and Graph Models and Application to Sonification.使用语言和图形模型进行端到端蛋白质正常模式频率预测及其在可听化中的应用。
ACS Nano. 2022 Dec 27;16(12):20656-20670. doi: 10.1021/acsnano.2c07681. Epub 2022 Nov 23.
9
CollagenTransformer: End-to-End Transformer Model to Predict Thermal Stability of Collagen Triple Helices Using an NLP Approach.胶原转换器:使用自然语言处理方法预测胶原三螺旋热稳定性的端到端转换器模型。
ACS Biomater Sci Eng. 2022 Oct 10;8(10):4301-4310. doi: 10.1021/acsbiomaterials.2c00737. Epub 2022 Sep 23.
10
End-to-End Deep Learning Model to Predict and Design Secondary Structure Content of Structural Proteins.端到端深度学习模型预测和设计结构蛋白的二级结构内容。
ACS Biomater Sci Eng. 2022 Mar 14;8(3):1156-1165. doi: 10.1021/acsbiomaterials.1c01343. Epub 2022 Feb 7.