• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于生成式预训练转换器模型和牙科医疗记录的医疗文本预测和建议。

Medical Text Prediction and Suggestion Using Generative Pretrained Transformer Models with Dental Medical Notes.

机构信息

The Abigail Wexner Research Institute, Nationwide Children's Hospital, Columbus, Ohio, United States.

Pediatric Dentistry, Nationwide Children's Hospital, Columbus, Ohio, United States.

出版信息

Methods Inf Med. 2022 Dec;61(5-06):195-200. doi: 10.1055/a-1900-7351. Epub 2022 Jul 14.

DOI:10.1055/a-1900-7351
PMID:35835447
Abstract

BACKGROUND

Generative pretrained transformer (GPT) models are one of the latest large pretrained natural language processing models that enables model training with limited datasets and reduces dependency on large datasets, which are scarce and costly to establish and maintain. There is a rising interest to explore the use of GPT models in health care.

OBJECTIVE

We investigate the performance of GPT-2 and GPT-Neo models for medical text prediction using 374,787 free-text dental notes.

METHODS

We fine-tune pretrained GPT-2 and GPT-Neo models for next word prediction on a dataset of over 374,000 manually written sections of dental clinical notes. Each model was trained on 80% of the dataset, validated on 10%, and tested on the remaining 10%. We report model performance in terms of next word prediction accuracy and loss. Additionally, we analyze the performance of the models on different types of prediction tokens for categories. For comparison, we also fine-tuned a non-GPT pretrained neural network model, XLNet (large), for next word prediction. We annotate each token in 100 randomly sampled notes by category (e.g., names, abbreviations, clinical terms, punctuation, etc.) and compare the performance of each model by token category.

RESULTS

Models present acceptable accuracy scores (GPT-2: 76%; GPT-Neo: 53%), and the GPT-2 model also performs better in manual evaluations, especially for names, abbreviations, and punctuation. Both GPT models outperformed XLNet in terms of accuracy. The results suggest that pretrained models have the potential to assist medical charting in the future. We share the lessons learned, insights, and suggestions for future implementations.

CONCLUSION

The results suggest that pretrained models have the potential to assist medical charting in the future. Our study presented one of the first implementations of the GPT model used with medical notes.

摘要

背景

生成式预训练转换器(GPT)模型是最新的大型预训练自然语言处理模型之一,它可以在数据集有限的情况下进行模型训练,并减少对稀缺且建立和维护成本高的大型数据集的依赖。人们越来越有兴趣探索在医疗保健中使用 GPT 模型。

目的

我们使用 374787 份自由文本牙科记录来研究 GPT-2 和 GPT-Neo 模型在医疗文本预测方面的性能。

方法

我们在超过 374000 份人工书写的牙科临床记录片段数据集上对预训练的 GPT-2 和 GPT-Neo 模型进行微调,以进行下一个单词预测。每个模型在数据集的 80%上进行训练,在 10%上进行验证,在剩余的 10%上进行测试。我们报告模型在预测准确性和损失方面的性能。此外,我们还分析了模型在不同类型的预测标记方面对类别的性能。为了比较,我们还对预训练的神经网络模型 XLNet(大型)进行了微调,用于下一个单词预测。我们按类别(例如,名称、缩写、临床术语、标点符号等)对 100 个随机抽样记录中的每个标记进行注释,并通过标记类别比较每个模型的性能。

结果

模型表现出可接受的准确率(GPT-2:76%;GPT-Neo:53%),并且 GPT-2 模型在人工评估中表现更好,特别是在名称、缩写和标点符号方面。在准确性方面,GPT 模型均优于 XLNet。结果表明,预训练模型具有在未来协助医疗记录的潜力。我们分享了经验教训、见解和对未来实施的建议。

结论

结果表明,预训练模型具有在未来协助医疗记录的潜力。我们的研究首次实现了将 GPT 模型应用于医疗记录。

相似文献

1
Medical Text Prediction and Suggestion Using Generative Pretrained Transformer Models with Dental Medical Notes.基于生成式预训练转换器模型和牙科医疗记录的医疗文本预测和建议。
Methods Inf Med. 2022 Dec;61(5-06):195-200. doi: 10.1055/a-1900-7351. Epub 2022 Jul 14.
2
A large language model-based generative natural language processing framework fine-tuned on clinical notes accurately extracts headache frequency from electronic health records.基于大型语言模型的生成式自然语言处理框架,在临床笔记上进行了微调,能够从电子健康记录中准确提取头痛频率。
Headache. 2024 Apr;64(4):400-409. doi: 10.1111/head.14702. Epub 2024 Mar 25.
3
A Large Language Model-Based Generative Natural Language Processing Framework Finetuned on Clinical Notes Accurately Extracts Headache Frequency from Electronic Health Records.一种基于大语言模型的生成式自然语言处理框架,在临床笔记上进行微调后,能准确从电子健康记录中提取头痛频率。
medRxiv. 2023 Oct 3:2023.10.02.23296403. doi: 10.1101/2023.10.02.23296403.
4
Large-scale identification of social and behavioral determinants of health from clinical notes: comparison of Latent Semantic Indexing and Generative Pretrained Transformer (GPT) models.从临床记录中大规模识别健康的社会和行为决定因素:潜在语义索引和生成式预训练转换器 (GPT) 模型的比较。
BMC Med Inform Decis Mak. 2024 Oct 10;24(1):296. doi: 10.1186/s12911-024-02705-x.
5
Operationalizing and Implementing Pretrained, Large Artificial Intelligence Linguistic Models in the US Health Care System: Outlook of Generative Pretrained Transformer 3 (GPT-3) as a Service Model.在美国医疗保健系统中实施和应用预训练的大型人工智能语言模型:生成式预训练变换器3(GPT-3)作为服务模型的前景
JMIR Med Inform. 2022 Feb 10;10(2):e32875. doi: 10.2196/32875.
6
Few-Shot Learning for Clinical Natural Language Processing Using Siamese Neural Networks: Algorithm Development and Validation Study.使用暹罗神经网络的临床自然语言处理少样本学习:算法开发与验证研究
JMIR AI. 2023 May 4;2:e44293. doi: 10.2196/44293.
7
Assessing the Ability of a Large Language Model to Score Free-Text Medical Student Clinical Notes: Quantitative Study.评估大型语言模型对自由文本医学生临床笔记评分的能力:定量研究。
JMIR Med Educ. 2024 Jul 25;10:e56342. doi: 10.2196/56342.
8
A comparative study on deep learning models for text classification of unstructured medical notes with various levels of class imbalance.深度学习模型在不同类别不平衡程度的非结构化医疗记录文本分类中的对比研究。
BMC Med Res Methodol. 2022 Jul 2;22(1):181. doi: 10.1186/s12874-022-01665-y.
9
Generative pretrained transformer-4, an artificial intelligence text predictive model, has a high capability for passing novel written radiology exam questions.生成式预训练转换器-4,一种人工智能文本预测模型,具有通过新型书面放射科考试问题的高能力。
Int J Comput Assist Radiol Surg. 2024 Apr;19(4):645-653. doi: 10.1007/s11548-024-03071-9. Epub 2024 Feb 21.
10
Generative Pre-trained Transformer (GPT) based model with relative attention for de novo drug design.基于生成式预训练转换器(GPT)的相对注意力模型在从头设计药物中的应用。
Comput Biol Chem. 2023 Oct;106:107911. doi: 10.1016/j.compbiolchem.2023.107911. Epub 2023 Jun 29.

引用本文的文献

1
The Development Landscape of Large Language Models for Biomedical Applications.用于生物医学应用的大语言模型的发展态势
Annu Rev Biomed Data Sci. 2025 Aug;8(1):251-274. doi: 10.1146/annurev-biodatasci-102224-074736. Epub 2025 Apr 1.
2
Artificial Intelligence in Dental Education: Opportunities and Challenges of Large Language Models and Multimodal Foundation Models.人工智能在牙科教育中的应用:大型语言模型和多模态基础模型的机遇与挑战。
JMIR Med Educ. 2024 Sep 27;10:e52346. doi: 10.2196/52346.
3
Viability of Open Large Language Models for Clinical Documentation in German Health Care: Real-World Model Evaluation Study.
德国医疗保健领域中开源大型语言模型用于临床文档记录的可行性:真实世界模型评估研究
JMIR Med Inform. 2024 Aug 28;12:e59617. doi: 10.2196/59617.
4
Integration and Assessment of ChatGPT in Medical Case Reporting: A Multifaceted Approach.ChatGPT在医学病例报告中的整合与评估:一种多维度方法
Eur J Investig Health Psychol Educ. 2024 Mar 30;14(4):888-901. doi: 10.3390/ejihpe14040057.
5
Year 2022 in Medical Natural Language Processing: Availability of Language Models as a Step in the Democratization of NLP in the Biomedical Area.2022 年医学自然语言处理:语言模型的可用性是生物医学领域 NLP 民主化的一步。
Yearb Med Inform. 2023 Aug;32(1):244-252. doi: 10.1055/s-0043-1768752. Epub 2023 Dec 26.
6
Extracting Medical Information From Free-Text and Unstructured Patient-Generated Health Data Using Natural Language Processing Methods: Feasibility Study With Real-world Data.使用自然语言处理方法从自由文本和非结构化患者生成的健康数据中提取医学信息:基于真实世界数据的可行性研究
JMIR Form Res. 2023 Mar 7;7:e43014. doi: 10.2196/43014.