• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

MechBERT:用于提取关于机械应力和应变的化学与性质关系的语言模型。

MechBERT: Language Models for Extracting Chemical and Property Relationships about Mechanical Stress and Strain.

作者信息

Kumar Pankaj, Kabra Saurabh, Cole Jacqueline M

机构信息

Cavendish Laboratory, Department of Physics, University of Cambridge, J. J. Thomson Avenue, Cambridge CB3 0HE. U.K.

ISIS Neutron and Muon Source, STFC Rutherford Appleton Laboratory, Harwell Science and Innovation Campus, Didcot OX11 0QX, U.K.

出版信息

J Chem Inf Model. 2025 Feb 24;65(4):1873-1888. doi: 10.1021/acs.jcim.4c00857. Epub 2025 Jan 31.

DOI:10.1021/acs.jcim.4c00857
PMID:39888859
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11863389/
Abstract

Language models are transforming materials-aware natural-language processing by enabling the extraction of dynamic, context-rich information from unstructured text, thus, moving beyond the limitations of traditional information-extraction methods. Moreover, small language models are on the rise because some of them can perform better than large language models (LLMs) when given domain-specific question-answer tasks, especially about an application area that relies on a highly specialized vernacular, such as materials science. We therefore present a new class of MechBERT language models for understanding mechanical stress and strain in materials. These employ Bidirectional Encoder Representations for transformer (BERT) architectures. We showcase four MechBERT models, all of which were pretrained on a corpus of documents that are textually rich in chemicals and their stress-strain properties and were fine-tuned on question-answering tasks. We evaluated the level of performance of our models on domain-specific as well as general English-language question-answer tasks and also explored the influence of the size and type of BERT architectures on model performance. We find that our MechBERT models outperform BERT-based models of the same size and maintain relevancy better than much larger BERT-based models when tasked with domain-specific question-answering tasks within the stress-strain engineering sector. These small language models also enable much faster processing and require a much smaller fraction of data to pretrain them, affording them greater operational efficiency and energy sustainability than LLMs.

摘要

语言模型正在改变材料感知自然语言处理,通过从非结构化文本中提取动态的、上下文丰富的信息,从而突破传统信息提取方法的局限性。此外,小型语言模型正在兴起,因为在给定特定领域的问答任务时,其中一些模型的表现优于大型语言模型(LLMs),特别是在依赖高度专业化术语的应用领域,如材料科学。因此,我们提出了一类新的MechBERT语言模型,用于理解材料中的机械应力和应变。这些模型采用了用于Transformer(BERT)架构的双向编码器表示。我们展示了四个MechBERT模型,所有这些模型都在一个文本丰富的化学物质及其应力 - 应变特性的文档语料库上进行了预训练,并在问答任务上进行了微调。我们评估了我们的模型在特定领域以及通用英语问答任务上的性能水平,并探讨了BERT架构的大小和类型对模型性能的影响。我们发现,在应力 - 应变工程领域的特定领域问答任务中,我们的MechBERT模型优于相同大小的基于BERT的模型,并且比大得多的基于BERT的模型更好地保持相关性。这些小型语言模型还能够实现更快的处理速度,并且在预训练时需要的数据量要少得多,与大型语言模型相比,它们具有更高的运营效率和能源可持续性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/d50f672367ae/ci4c00857_0013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/50f2124372ab/ci4c00857_0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/b14e2bdfa347/ci4c00857_0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/2a94d0af90b3/ci4c00857_0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/99cc1b4cd4b1/ci4c00857_0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/1f133a7a9e1e/ci4c00857_0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/3e904e16b9ef/ci4c00857_0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/5adf3e1236d6/ci4c00857_0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/4d10a70fd0cf/ci4c00857_0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/b93cf01c1bf9/ci4c00857_0009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/60dd2b42d5e9/ci4c00857_0010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/56915c084214/ci4c00857_0011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/59c500c30094/ci4c00857_0012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/d50f672367ae/ci4c00857_0013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/50f2124372ab/ci4c00857_0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/b14e2bdfa347/ci4c00857_0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/2a94d0af90b3/ci4c00857_0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/99cc1b4cd4b1/ci4c00857_0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/1f133a7a9e1e/ci4c00857_0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/3e904e16b9ef/ci4c00857_0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/5adf3e1236d6/ci4c00857_0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/4d10a70fd0cf/ci4c00857_0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/b93cf01c1bf9/ci4c00857_0009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/60dd2b42d5e9/ci4c00857_0010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/56915c084214/ci4c00857_0011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/59c500c30094/ci4c00857_0012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c16b/11863389/d50f672367ae/ci4c00857_0013.jpg

相似文献

1
MechBERT: Language Models for Extracting Chemical and Property Relationships about Mechanical Stress and Strain.MechBERT:用于提取关于机械应力和应变的化学与性质关系的语言模型。
J Chem Inf Model. 2025 Feb 24;65(4):1873-1888. doi: 10.1021/acs.jcim.4c00857. Epub 2025 Jan 31.
2
Oversampling effect in pretraining for bidirectional encoder representations from transformers (BERT) to localize medical BERT and enhance biomedical BERT.在基于转换器的双向编码器表示预训练(BERT)中进行过采样,以定位医学 BERT 并增强生物医学 BERT。
Artif Intell Med. 2024 Jul;153:102889. doi: 10.1016/j.artmed.2024.102889. Epub 2024 May 5.
3
Extracting comprehensive clinical information for breast cancer using deep learning methods.利用深度学习方法提取乳腺癌全面临床信息。
Int J Med Inform. 2019 Dec;132:103985. doi: 10.1016/j.ijmedinf.2019.103985. Epub 2019 Oct 2.
4
Multifaceted Natural Language Processing Task-Based Evaluation of Bidirectional Encoder Representations From Transformers Models for Bilingual (Korean and English) Clinical Notes: Algorithm Development and Validation.基于转换器模型的双向编码器表示的多方面自然语言处理任务评估在双语(韩语和英语)临床笔记中的应用:算法开发和验证。
JMIR Med Inform. 2024 Oct 30;12:e52897. doi: 10.2196/52897.
5
BatteryBERT: A Pretrained Language Model for Battery Database Enhancement.电池 BERT:用于电池数据库增强的预训练语言模型。
J Chem Inf Model. 2022 Dec 26;62(24):6365-6377. doi: 10.1021/acs.jcim.2c00035. Epub 2022 May 9.
6
When BERT meets Bilbo: a learning curve analysis of pretrained language model on disease classification.当 BERT 遇见比尔博:预训练语言模型在疾病分类上的学习曲线分析。
BMC Med Inform Decis Mak. 2022 Apr 5;21(Suppl 9):377. doi: 10.1186/s12911-022-01829-2.
7
Positional embeddings and zero-shot learning using BERT for molecular-property prediction.使用BERT进行位置嵌入和零样本学习以预测分子性质
J Cheminform. 2025 Feb 5;17(1):17. doi: 10.1186/s13321-025-00959-9.
8
CACER: Clinical concept Annotations for Cancer Events and Relations.CACER:癌症事件与关系的临床概念注释。
J Am Med Inform Assoc. 2024 Nov 1;31(11):2583-2594. doi: 10.1093/jamia/ocae231.
9
A Question-and-Answer System to Extract Data From Free-Text Oncological Pathology Reports (CancerBERT Network): Development Study.从自由文本肿瘤病理学报告(CancerBERT 网络)中提取数据的问答系统:开发研究。
J Med Internet Res. 2022 Mar 23;24(3):e27210. doi: 10.2196/27210.
10
OpticalBERT and OpticalTable-SQA: Text- and Table-Based Language Models for the Optical-Materials Domain.光学 BERT 和光学 Table-SQA:面向光学材料领域的基于文本和表格的语言模型。
J Chem Inf Model. 2023 Apr 10;63(7):1961-1981. doi: 10.1021/acs.jcim.2c01259. Epub 2023 Mar 20.

本文引用的文献

1
A database of thermally activated delayed fluorescent molecules auto-generated from scientific literature with ChemDataExtractor.利用 ChemDataExtractor 从科学文献中自动生成热激活延迟荧光分子数据库。
Sci Data. 2024 Jan 17;11(1):80. doi: 10.1038/s41597-023-02897-3.
2
Automated Construction of a Photocatalysis Dataset for Water-Splitting Applications.自动化构建用于水分解应用的光催化数据集。
Sci Data. 2023 Sep 22;10(1):651. doi: 10.1038/s41597-023-02511-6.
3
OpticalBERT and OpticalTable-SQA: Text- and Table-Based Language Models for the Optical-Materials Domain.
光学 BERT 和光学 Table-SQA:面向光学材料领域的基于文本和表格的语言模型。
J Chem Inf Model. 2023 Apr 10;63(7):1961-1981. doi: 10.1021/acs.jcim.2c01259. Epub 2023 Mar 20.
4
A thermoelectric materials database auto-generated from the scientific literature using ChemDataExtractor.使用 ChemDataExtractor 从科学文献中自动生成的热电材料数据库。
Sci Data. 2022 Oct 22;9(1):648. doi: 10.1038/s41597-022-01752-1.
5
BatteryBERT: A Pretrained Language Model for Battery Database Enhancement.电池 BERT:用于电池数据库增强的预训练语言模型。
J Chem Inf Model. 2022 Dec 26;62(24):6365-6377. doi: 10.1021/acs.jcim.2c00035. Epub 2022 May 9.
6
A database of refractive indices and dielectric constants auto-generated using ChemDataExtractor.使用 ChemDataExtractor 自动生成的折射率和介电常数数据库。
Sci Data. 2022 May 3;9(1):192. doi: 10.1038/s41597-022-01295-5.
7
Auto-generated database of semiconductor band gaps using ChemDataExtractor.使用 ChemDataExtractor 自动生成半导体带隙数据库。
Sci Data. 2022 May 3;9(1):193. doi: 10.1038/s41597-022-01294-6.
8
PDFDataExtractor: A Tool for Reading Scientific Text and Interpreting Metadata from the Typeset Literature in the Portable Document Format.PDFDataExtractor:一种从可移植文档格式中的排版文献中读取科学文本和解释元数据的工具。
J Chem Inf Model. 2022 Apr 11;62(7):1633-1643. doi: 10.1021/acs.jcim.1c01198. Epub 2022 Mar 29.
9
Single Model for Organic and Inorganic Chemical Named Entity Recognition in ChemDataExtractor.在 ChemDataExtractor 中进行有机和无机化学命名实体识别的单一模型。
J Chem Inf Model. 2022 Mar 14;62(5):1207-1213. doi: 10.1021/acs.jcim.1c01199. Epub 2022 Feb 24.
10
ChemDataExtractor 2.0: Autopopulated Ontologies for Materials Science.ChemDataExtractor 2.0:材料科学自动填充本体。
J Chem Inf Model. 2021 Sep 27;61(9):4280-4289. doi: 10.1021/acs.jcim.1c00446. Epub 2021 Sep 16.