• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

根据基于能力的医学教育课程评估ChatGPT回答微生物学一阶和二阶知识问题的能力。

Assessing the Capability of ChatGPT in Answering First- and Second-Order Knowledge Questions on Microbiology as per Competency-Based Medical Education Curriculum.

作者信息

Das Dipmala, Kumar Nikhil, Longjam Langamba Angom, Sinha Ranwir, Deb Roy Asitava, Mondal Himel, Gupta Pratima

机构信息

Microbiology, All India Institute of Medical Sciences, Deoghar, Deoghar, IND.

Pathology, All India Institute of Medical Sciences, Deoghar, Deoghar, IND.

出版信息

Cureus. 2023 Mar 12;15(3):e36034. doi: 10.7759/cureus.36034. eCollection 2023 Mar.

DOI:10.7759/cureus.36034
PMID:37056538
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10086829/
Abstract

Background and objective ChatGPT is an artificial intelligence (AI) language model that has been trained to process and respond to questions across a wide range of topics. It is also capable of solving problems in medical educational topics. However, the capability of ChatGPT to accurately answer first- and second-order knowledge questions in the field of microbiology has not been explored so far. Hence, in this study, we aimed to analyze the capability of ChatGPT in answering first- and second-order questions on the subject of microbiology. Materials and methods Based on the competency-based medical education (CBME) curriculum of the subject of microbiology, we prepared a set of first-order and second-order questions. For the total of eight modules in the CBME curriculum for microbiology, we prepared six first-order and six second-order knowledge questions according to the National Medical Commission-recommended CBME curriculum, amounting to a total of (8 x 12) 96 questions. The questions were checked for content validity by three expert microbiologists. These questions were used to converse with ChatGPT by a single user and responses were recorded for further analysis. The answers were scored by three microbiologists on a rating scale of 0-5. The average of three scores was taken as the final score for analysis. As the data were not normally distributed, we used a non-parametric statistical test. The overall scores were tested by a one-sample median test with hypothetical values of 4 and 5. The scores of answers to first-order and second-order questions were compared by the Mann-Whitney U test. Module-wise responses were tested by the Kruskall-Wallis test followed by the post hoc test for pairwise comparisons. Results The overall score of 96 answers was 4.04 ±0.37 (median: 4.17, Q1-Q3: 3.88-4.33) with the mean score of answers to first-order knowledge questions being 4.07 ±0.32 (median: 4.17, Q1-Q3: 4-4.33) and that of answers to second-order knowledge questions being 3.99 ±0.43 (median: 4, Q1-Q3: 3.67-4.33) (Mann-Whitney p=0.4). The score was significantly below the score of 5 (one-sample median test p<0.0001) and similar to 4 (one-sample median test p=0.09). Overall, there was a variation in median scores obtained in eight categories of topics in microbiology, indicating inconsistent performance in different topics. Conclusion The results of the study indicate that ChatGPT is capable of answering both first- and second-order knowledge questions related to the subject of microbiology. The model achieved an accuracy of approximately 80% and there was no difference between the model's capability of answering first-order questions and second-order knowledge questions. The findings of this study suggest that ChatGPT has the potential to be an effective tool for automated question-answering in the field of microbiology. However, continued improvement in the training and development of language models is necessary to enhance their performance and make them suitable for academic use.

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8013/10086829/97e1460b3340/cureus-0015-00000036034-i05.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8013/10086829/0a5fdad60ae9/cureus-0015-00000036034-i01.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8013/10086829/aef09e8af3ef/cureus-0015-00000036034-i02.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8013/10086829/a506fa950f46/cureus-0015-00000036034-i03.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8013/10086829/4286cb36edb9/cureus-0015-00000036034-i04.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8013/10086829/97e1460b3340/cureus-0015-00000036034-i05.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8013/10086829/0a5fdad60ae9/cureus-0015-00000036034-i01.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8013/10086829/aef09e8af3ef/cureus-0015-00000036034-i02.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8013/10086829/a506fa950f46/cureus-0015-00000036034-i03.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8013/10086829/4286cb36edb9/cureus-0015-00000036034-i04.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8013/10086829/97e1460b3340/cureus-0015-00000036034-i05.jpg
摘要

背景与目的 ChatGPT是一种人工智能(AI)语言模型,经过训练可处理和回答广泛主题的问题。它也能够解决医学教育主题中的问题。然而,ChatGPT在微生物学领域准确回答一阶和二阶知识问题的能力迄今尚未得到探索。因此,在本研究中,我们旨在分析ChatGPT回答微生物学主题一阶和二阶问题的能力。

材料与方法 基于微生物学主题的基于胜任力的医学教育(CBME)课程,我们准备了一组一阶和二阶问题。针对微生物学CBME课程的总共八个模块,我们根据国家医学委员会推荐的CBME课程准备了六个一阶和六个二阶知识问题,共计(8×12)96个问题。这些问题由三位微生物学专家检查内容效度。这些问题由一名用户用于与ChatGPT进行对话,并记录回答以供进一步分析。答案由三位微生物学家按照0至5的评分量表进行评分。取三个分数的平均值作为最终分析分数。由于数据非正态分布,我们使用非参数统计检验。总体分数通过单样本中位数检验进行测试,假设值为4和5。一阶和二阶问题答案的分数通过曼-惠特尼U检验进行比较。按模块的回答通过Kruskal-Wallis检验进行测试,随后进行两两比较的事后检验。

结果 96个答案的总体分数为4.04±0.37(中位数:4.17,四分位数间距:3.88 - 4.33),一阶知识问题答案的平均分数为4.07±0.32(中位数:4.17,四分位数间距:4 - 4.33),二阶知识问题答案的平均分数为3.99±0.43(中位数:4,四分位数间距:3.67 - 4.33)(曼-惠特尼检验p = 0.4)。该分数显著低于5分(单样本中位数检验p<0.0001)且与4分相似(单样本中位数检验p = 0.09)。总体而言,微生物学八个主题类别获得的中位数分数存在差异,表明在不同主题中的表现不一致。

结论 研究结果表明,ChatGPT能够回答与微生物学主题相关的一阶和二阶知识问题。该模型的准确率约为80%,且在回答一阶问题和二阶知识问题的能力之间没有差异。本研究结果表明,ChatGPT有潜力成为微生物学领域自动问答的有效工具。然而,有必要持续改进语言模型的训练和开发,以提高其性能并使其适用于学术用途。

相似文献

1
Assessing the Capability of ChatGPT in Answering First- and Second-Order Knowledge Questions on Microbiology as per Competency-Based Medical Education Curriculum.根据基于能力的医学教育课程评估ChatGPT回答微生物学一阶和二阶知识问题的能力。
Cureus. 2023 Mar 12;15(3):e36034. doi: 10.7759/cureus.36034. eCollection 2023 Mar.
2
Evaluating ChatGPT's Ability to Solve Higher-Order Questions on the Competency-Based Medical Education Curriculum in Medical Biochemistry.评估ChatGPT解决医学基础生物化学基于能力的医学教育课程中高阶问题的能力。
Cureus. 2023 Apr 2;15(4):e37023. doi: 10.7759/cureus.37023. eCollection 2023 Apr.
3
Applicability of ChatGPT in Assisting to Solve Higher Order Problems in Pathology.ChatGPT在协助解决病理学高阶问题中的适用性。
Cureus. 2023 Feb 20;15(2):e35237. doi: 10.7759/cureus.35237. eCollection 2023 Feb.
4
Large Language Models in Hematology Case Solving: A Comparative Study of ChatGPT-3.5, Google Bard, and Microsoft Bing.大语言模型在血液学病例解决中的应用:ChatGPT-3.5、谷歌巴德和微软必应的比较研究
Cureus. 2023 Aug 21;15(8):e43861. doi: 10.7759/cureus.43861. eCollection 2023 Aug.
5
Analysing the Applicability of ChatGPT, Bard, and Bing to Generate Reasoning-Based Multiple-Choice Questions in Medical Physiology.分析ChatGPT、Bard和必应在医学生理学中生成基于推理的多项选择题的适用性。
Cureus. 2023 Jun 26;15(6):e40977. doi: 10.7759/cureus.40977. eCollection 2023 Jun.
6
Performance of ChatGPT on the Chinese Postgraduate Examination for Clinical Medicine: Survey Study.ChatGPT 在临床医学研究生入学考试中的表现:调查研究。
JMIR Med Educ. 2024 Feb 9;10:e48514. doi: 10.2196/48514.
7
Assessing the Accuracy of Information on Medication Abortion: A Comparative Analysis of ChatGPT and Google Bard AI.评估药物流产信息的准确性:ChatGPT与谷歌巴德人工智能的比较分析
Cureus. 2024 Jan 2;16(1):e51544. doi: 10.7759/cureus.51544. eCollection 2024 Jan.
8
Efficacy of ChatGPT in solving attitude, ethics, and communication case scenario used for competency-based medical education in India: A case study.ChatGPT在解决印度基于能力的医学教育中使用的态度、伦理和沟通案例场景方面的有效性:一项案例研究。
J Educ Health Promot. 2024 Feb 7;13:22. doi: 10.4103/jehp.jehp_625_23. eCollection 2024.
9
How Does ChatGPT Perform on the United States Medical Licensing Examination (USMLE)? The Implications of Large Language Models for Medical Education and Knowledge Assessment.ChatGPT在美国医师执照考试(USMLE)中的表现如何?大语言模型对医学教育和知识评估的影响。
JMIR Med Educ. 2023 Feb 8;9:e45312. doi: 10.2196/45312.
10
Evaluating ChatGPT to test its robustness as an interactive information database of radiation oncology and to assess its responses to common queries from radiotherapy patients: A single institution investigation.评估ChatGPT以测试其作为放射肿瘤学交互式信息数据库的稳健性,并评估其对放疗患者常见问题的回答:一项单机构调查。
Cancer Radiother. 2024 Jun;28(3):258-264. doi: 10.1016/j.canrad.2023.11.005. Epub 2024 Jun 12.

引用本文的文献

1
Assessing the Accuracy and Completeness of AI-Generated Dental Responses: An Evaluation of the Chat-GPT Model.评估人工智能生成的牙科回复的准确性和完整性:Chat-GPT模型的评估
Healthcare (Basel). 2025 Aug 28;13(17):2144. doi: 10.3390/healthcare13172144.
2
Identification and Categorization of the Top 100 Articles and the Future of Large Language Models: Thematic Analysis Using Bibliometric Analysis.100篇顶级文章的识别与分类以及大语言模型的未来:基于文献计量分析的主题分析
JMIR AI. 2025 Aug 27;4:e68603. doi: 10.2196/68603.
3
GastroGPT: Development and controlled testing of a proof-of-concept customized clinical language model.

本文引用的文献

1
Applicability of ChatGPT in Assisting to Solve Higher Order Problems in Pathology.ChatGPT在协助解决病理学高阶问题中的适用性。
Cureus. 2023 Feb 20;15(2):e35237. doi: 10.7759/cureus.35237. eCollection 2023 Feb.
2
Performance of ChatGPT on USMLE: Potential for AI-assisted medical education using large language models.ChatGPT在美国医师执照考试中的表现:使用大语言模型进行人工智能辅助医学教育的潜力。
PLOS Digit Health. 2023 Feb 9;2(2):e0000198. doi: 10.1371/journal.pdig.0000198. eCollection 2023 Feb.
3
Artificial Hallucinations in ChatGPT: Implications in Scientific Writing.
胃语大模型:一种概念验证型定制临床语言模型的开发与对照测试
Endosc Int Open. 2025 Aug 6;13:a26372163. doi: 10.1055/a-2637-2163. eCollection 2025.
4
Evaluating the Use of ChatGPT 3.5 and Bard as Self-Assessment Tools for Short Answer Questions in Undergraduate Ophthalmology.评估ChatGPT 3.5和Bard作为本科眼科简答题自我评估工具的使用情况。
Cureus. 2025 Jun 18;17(6):e86288. doi: 10.7759/cureus.86288. eCollection 2025 Jun.
5
The Diagnostic Performance of Large Language Models and Oral Medicine Consultants for Identifying Oral Lesions in Text-Based Clinical Scenarios: Prospective Comparative Study.大语言模型与口腔医学顾问在基于文本的临床场景中识别口腔病变的诊断性能:前瞻性比较研究
JMIR AI. 2025 Apr 24;4:e70566. doi: 10.2196/70566.
6
ChatGPT and Other Large Language Models in Medical Education - Scoping Literature Review.医学教育中的ChatGPT及其他大语言模型——文献综述
Med Sci Educ. 2024 Nov 13;35(1):555-567. doi: 10.1007/s40670-024-02206-6. eCollection 2025 Feb.
7
Chat GPT, Gemini or Meta AI: A comparison of AI platforms as a tool for answering higher-order questions in microbiology.Chat GPT、Gemini 还是 Meta AI:人工智能平台作为回答微生物学高阶问题工具的比较
J Postgrad Med. 2025 Jan 1;71(1):28-32. doi: 10.4103/jpgm.jpgm_775_24. Epub 2025 Mar 19.
8
Can ChatGPT pass the Turkish Orthopedics and Traumatology Board Examination? Turkish orthopedic surgeons versus artificial intelligence.ChatGPT能通过土耳其骨科学与创伤外科学委员会考试吗?土耳其骨科医生与人工智能的较量。
Ulus Travma Acil Cerrahi Derg. 2025 Mar;31(3):310-315. doi: 10.14744/tjtes.2025.07724.
9
Applications of Artificial Intelligence in Medical Education: A Systematic Review.人工智能在医学教育中的应用:一项系统综述。
Cureus. 2025 Mar 1;17(3):e79878. doi: 10.7759/cureus.79878. eCollection 2025 Mar.
10
Perceptions and Earliest Experiences of Medical Students and Faculty With ChatGPT in Medical Education: Qualitative Study.医学生和教师在医学教育中对ChatGPT的认知与早期体验:定性研究
JMIR Med Educ. 2025 Feb 20;11:e63400. doi: 10.2196/63400.
ChatGPT中的人工幻觉:对科学写作的影响
Cureus. 2023 Feb 19;15(2):e35179. doi: 10.7759/cureus.35179. eCollection 2023 Feb.
4
The future of medical education and research: Is ChatGPT a blessing or blight in disguise?医学教育与研究的未来:ChatGPT 是伪装的福祉还是祸根?
Med Educ Online. 2023 Dec;28(1):2181052. doi: 10.1080/10872981.2023.2181052.
5
Are ChatGPT’s knowledge and interpretation ability comparable to those of medical students in Korea for taking a parasitology examination?: a descriptive study.ChatGPT 的知识和解释能力与韩国医学生在寄生虫学考试中的表现相当吗?一项描述性研究。
J Educ Eval Health Prof. 2023;20:1. doi: 10.3352/jeehp.2023.20.1. Epub 2023 Jan 11.
6
Conduct Common Statistical Tests Online.在线进行常见统计测试。
Indian Dermatol Online J. 2022 Jun 24;13(4):539-542. doi: 10.4103/idoj.idoj_605_21. eCollection 2022 Jul-Aug.
7
A pilot study on case-based learning (CBL) in medical microbiology; students perspective.医学微生物学中基于案例学习(CBL)的初步研究;学生视角
Med J Armed Forces India. 2021 Feb;77(Suppl 1):S215-S219. doi: 10.1016/j.mjafi.2021.01.005. Epub 2021 Feb 2.
8
Self-directed learning: assessment of students' abilities and their perspective.自主学习:学生能力及其观点的评估
Adv Physiol Educ. 2020 Sep 1;44(3):383-386. doi: 10.1152/advan.00010.2020.
9
Capturing the Patient's Perspective: a Review of Advances in Natural Language Processing of Health-Related Text.捕捉患者视角:健康相关文本自然语言处理进展综述
Yearb Med Inform. 2017 Aug;26(1):214-227. doi: 10.15265/IY-2017-029. Epub 2017 Sep 11.
10
Competency-based medical education: An overview and application in pharmacology.基于能力的医学教育:概述及其在药理学中的应用
Indian J Pharmacol. 2016 Oct;48(Suppl 1):S5-S9. doi: 10.4103/0253-7613.193312.