• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

中文医学命名实体识别中大型语言模型的比较分析

Comparative Analysis of Large Language Models in Chinese Medical Named Entity Recognition.

作者信息

Zhu Zhichao, Zhao Qing, Li Jianjiang, Ge Yanhu, Ding Xingjian, Gu Tao, Zou Jingchen, Lv Sirui, Wang Sheng, Yang Ji-Jiang

机构信息

College of Computer Science, Beijing University of Technology, Beijing 100124, China.

Department of Anesthesiology, Beijing Anzhen Hospital, Capital Medical University, Beijing 100013, China.

出版信息

Bioengineering (Basel). 2024 Sep 29;11(10):982. doi: 10.3390/bioengineering11100982.

DOI:10.3390/bioengineering11100982
PMID:39451358
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11504658/
Abstract

The emergence of large language models (LLMs) has provided robust support for application tasks across various domains, such as name entity recognition (NER) in the general domain. However, due to the particularity of the medical domain, the research on understanding and improving the effectiveness of LLMs on biomedical named entity recognition (BNER) tasks remains relatively limited, especially in the context of Chinese text. In this study, we extensively evaluate several typical LLMs, including ChatGLM2-6B, GLM-130B, GPT-3.5, and GPT-4, on the Chinese BNER task by leveraging a real-world Chinese electronic medical record (EMR) dataset and a public dataset. The experimental results demonstrate the promising yet limited performance of LLMs with zero-shot and few-shot prompt designs for Chinese BNER tasks. More importantly, instruction fine-tuning significantly enhances the performance of LLMs. The fine-tuned offline ChatGLM2-6B surpassed the performance of the task-specific model BiLSTM+CRF (BC) on the real-world dataset. The best fine-tuned model, GPT-3.5, outperforms all other LLMs on the publicly available CCKS2017 dataset, even surpassing half of the baselines; however, it still remains challenging for it to surpass the state-of-the-art task-specific models, i.e., Dictionary-guided Attention Network (DGAN). To our knowledge, this study is the first attempt to evaluate the performance of LLMs on Chinese BNER tasks, which emphasizes the prospective and transformative implications of utilizing LLMs on Chinese BNER tasks. Furthermore, we summarize our findings into a set of actionable guidelines for future researchers on how to effectively leverage LLMs to become experts in specific tasks.

摘要

大语言模型(LLMs)的出现为跨领域的应用任务提供了强大支持,比如通用领域的命名实体识别(NER)。然而,由于医学领域的特殊性,关于理解和提高大语言模型在生物医学命名实体识别(BNER)任务上的有效性的研究仍然相对有限,尤其是在中文文本的背景下。在本研究中,我们通过利用真实世界的中文电子病历(EMR)数据集和一个公共数据集,对几个典型的大语言模型进行了广泛评估,包括ChatGLM2 - 6B、GLM - 130B、GPT - 3.5和GPT - 4,用于中文BNER任务。实验结果表明,对于中文BNER任务,大语言模型在零样本和少样本提示设计下的表现有前景但有限。更重要的是,指令微调显著提高了大语言模型的性能。经过离线微调的ChatGLM2 - 6B在真实世界数据集上超过了特定任务模型BiLSTM + CRF(BC)的性能。最佳微调模型GPT - 3.5在公开可用的CCKS2017数据集上优于所有其他大语言模型,甚至超过了一半的基线模型;然而,要超越当前最先进的特定任务模型,即字典引导注意力网络(DGAN),对它来说仍然具有挑战性。据我们所知,本研究是首次评估大语言模型在中文BNER任务上的性能,强调了在中文BNER任务中利用大语言模型的前瞻性和变革性影响。此外,我们将研究结果总结为一套可操作的指南,供未来的研究人员参考,以了解如何有效利用大语言模型成为特定任务的专家。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/488e/11504658/4c4c40494b0f/bioengineering-11-00982-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/488e/11504658/909922cdc54a/bioengineering-11-00982-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/488e/11504658/4c4c40494b0f/bioengineering-11-00982-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/488e/11504658/909922cdc54a/bioengineering-11-00982-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/488e/11504658/4c4c40494b0f/bioengineering-11-00982-g002.jpg

相似文献

1
Comparative Analysis of Large Language Models in Chinese Medical Named Entity Recognition.中文医学命名实体识别中大型语言模型的比较分析
Bioengineering (Basel). 2024 Sep 29;11(10):982. doi: 10.3390/bioengineering11100982.
2
Improving large language models for clinical named entity recognition via prompt engineering.通过提示工程改进临床命名实体识别的大型语言模型。
J Am Med Inform Assoc. 2024 Sep 1;31(9):1812-1820. doi: 10.1093/jamia/ocad259.
3
Advancing entity recognition in biomedicine via instruction tuning of large language models.通过指令调整大型语言模型推进生物医学中的实体识别。
Bioinformatics. 2024 Mar 29;40(4). doi: 10.1093/bioinformatics/btae163.
4
Sample Size Considerations for Fine-Tuning Large Language Models for Named Entity Recognition Tasks: Methodological Study.用于命名实体识别任务的大语言模型微调的样本量考量:方法学研究
JMIR AI. 2024 May 16;3:e52095. doi: 10.2196/52095.
5
BioInstruct: instruction tuning of large language models for biomedical natural language processing.BioInstruct:用于生物医学自然语言处理的大型语言模型的指令调整。
J Am Med Inform Assoc. 2024 Sep 1;31(9):1821-1832. doi: 10.1093/jamia/ocae122.
6
An Empirical Evaluation of Prompting Strategies for Large Language Models in Zero-Shot Clinical Natural Language Processing: Algorithm Development and Validation Study.零样本临床自然语言处理中大型语言模型提示策略的实证评估:算法开发与验证研究
JMIR Med Inform. 2024 Apr 8;12:e55318. doi: 10.2196/55318.
7
Taiyi: a bilingual fine-tuned large language model for diverse biomedical tasks.太乙:一个用于多种生物医学任务的双语精调大型语言模型。
J Am Med Inform Assoc. 2024 Sep 1;31(9):1865-1874. doi: 10.1093/jamia/ocae037.
8
Benchmarking Large Language Models in Evidence-Based Medicine.基于循证医学的大型语言模型基准测试。
IEEE J Biomed Health Inform. 2024 Oct 21;PP. doi: 10.1109/JBHI.2024.3483816.
9
Learning to Make Rare and Complex Diagnoses With Generative AI Assistance: Qualitative Study of Popular Large Language Models.利用生成式人工智能辅助学习罕见且复杂的诊断:对流行的大型语言模型的定性研究。
JMIR Med Educ. 2024 Feb 13;10:e51391. doi: 10.2196/51391.
10
Investigating the Impact of Prompt Engineering on the Performance of Large Language Models for Standardizing Obstetric Diagnosis Text: Comparative Study.探究提示工程对用于产科诊断文本标准化的大语言模型性能的影响:比较研究
JMIR Form Res. 2024 Feb 8;8:e53216. doi: 10.2196/53216.

引用本文的文献

1
Automated Extraction of Key Entities from Non-English Mammography Reports Using Named Entity Recognition with Prompt Engineering.使用带有提示工程的命名实体识别从非英语乳腺钼靶报告中自动提取关键实体
Bioengineering (Basel). 2025 Feb 10;12(2):168. doi: 10.3390/bioengineering12020168.

本文引用的文献

1
Leveraging Rule-Based NLP to Translate Textual Reports as Structured Inputs Automatically Processed by a Clinical Decision Support System.利用基于规则的自然语言处理技术自动将文本报告转换为临床决策支持系统可处理的结构化输入。
Stud Health Technol Inform. 2024 Aug 22;316:1861-1865. doi: 10.3233/SHTI240794.
2
Knowledge Guided Feature Aggregation for the Prediction of Chronic Obstructive Pulmonary Disease With Chinese EMRs.基于知识引导的特征聚合预测中国电子病历中的慢性阻塞性肺疾病
IEEE/ACM Trans Comput Biol Bioinform. 2023 Nov-Dec;20(6):3343-3352. doi: 10.1109/TCBB.2022.3198798. Epub 2023 Dec 25.
3
SiBERT: A Siamese-based BERT network for Chinese medical entities alignment.
SiBERT:一种基于 Siamese 的 BERT 网络,用于中文医疗实体对齐。
Methods. 2022 Sep;205:133-139. doi: 10.1016/j.ymeth.2022.07.003. Epub 2022 Jul 4.
4
Swarm Learning for decentralized and confidential clinical machine learning.群体学习用于去中心化和保密的临床机器学习。
Nature. 2021 Jun;594(7862):265-270. doi: 10.1038/s41586-021-03583-3. Epub 2021 May 26.
5
Chinese clinical named entity recognition with variant neural structures based on BERT methods.基于 BERT 方法的中文临床命名实体识别与变体神经结构。
J Biomed Inform. 2020 Jul;107:103422. doi: 10.1016/j.jbi.2020.103422. Epub 2020 Apr 28.
6
An attention-based deep learning model for clinical named entity recognition of Chinese electronic medical records.基于注意力的深度学习模型在中文电子病历临床命名实体识别中的应用。
BMC Med Inform Decis Mak. 2019 Dec 5;19(Suppl 5):235. doi: 10.1186/s12911-019-0933-6.
7
Document-level attention-based BiLSTM-CRF incorporating disease dictionary for disease named entity recognition.基于文档级注意力的 BiLSTM-CRF 结合疾病词典的疾病命名实体识别。
Comput Biol Med. 2019 May;108:122-132. doi: 10.1016/j.compbiomed.2019.04.002. Epub 2019 Apr 7.
8
Recurrent neural networks with specialized word embeddings for health-domain named-entity recognition.用于健康领域命名实体识别的具有专用词嵌入的递归神经网络。
J Biomed Inform. 2017 Dec;76:102-109. doi: 10.1016/j.jbi.2017.11.007. Epub 2017 Nov 13.