Suppr超能文献

评估大型语言模型在摘要筛选中的有效性:一项对比分析。

Evaluating the effectiveness of large language models in abstract screening: a comparative analysis.

机构信息

Department of Biostatistics, University of North Carolina at Chapel Hill, Chapel Hill, NC, 27599, USA.

Department of Mathematics and Statistics, University of North Carolina at Greensboro, Greensboro, NC, 27402, USA.

出版信息

Syst Rev. 2024 Aug 21;13(1):219. doi: 10.1186/s13643-024-02609-x.

Abstract

OBJECTIVE

This study aimed to evaluate the performance of large language models (LLMs) in the task of abstract screening in systematic review and meta-analysis studies, exploring their effectiveness, efficiency, and potential integration into existing human expert-based workflows.

METHODS

We developed automation scripts in Python to interact with the APIs of several LLM tools, including ChatGPT v4.0, ChatGPT v3.5, Google PaLM 2, and Meta Llama 2, and latest tools including ChatGPT v4.0 turbo, ChatGPT v3.5 turbo, Google Gemini 1.0 pro, Meta Llama 3, and Claude 3. This study focused on three databases of abstracts and used them as benchmarks to evaluate the performance of these LLM tools in terms of sensitivity, specificity, and overall accuracy. The results of the LLM tools were compared to human-curated inclusion decisions, gold standard for systematic review and meta-analysis studies.

RESULTS

Different LLM tools had varying abilities in abstract screening. Chat GPT v4.0 demonstrated remarkable performance, with balanced sensitivity and specificity, and overall accuracy consistently reaching or exceeding 90%, indicating a high potential for LLMs in abstract screening tasks. The study found that LLMs could provide reliable results with minimal human effort and thus serve as a cost-effective and efficient alternative to traditional abstract screening methods.

CONCLUSION

While LLM tools are not yet ready to completely replace human experts in abstract screening, they show great promise in revolutionizing the process. They can serve as autonomous AI reviewers, contribute to collaborative workflows with human experts, and integrate with hybrid approaches to develop custom tools for increased efficiency. As technology continues to advance, LLMs are poised to play an increasingly important role in abstract screening, reshaping the workflow of systematic review and meta-analysis studies.

摘要

目的

本研究旨在评估大型语言模型(LLM)在系统评价和荟萃分析研究的摘要筛选任务中的性能,探索其有效性、效率以及潜在的整合到现有的基于人类专家的工作流程中。

方法

我们使用 Python 开发了自动化脚本,与多个 LLM 工具的 API 进行交互,包括 ChatGPT v4.0、ChatGPT v3.5、Google PaLM 2 和 Meta Lama 2,以及最新的工具包括 ChatGPT v4.0 turbo、ChatGPT v3.5 turbo、Google Gemini 1.0 pro、Meta Lama 3 和 Claude 3。本研究主要关注三个摘要数据库,并将其用作基准,以评估这些 LLM 工具在敏感性、特异性和整体准确性方面的性能。LLM 工具的结果与人工筛选的纳入决策、系统评价和荟萃分析研究的黄金标准进行了比较。

结果

不同的 LLM 工具在摘要筛选方面具有不同的能力。Chat GPT v4.0 表现出出色的性能,具有平衡的敏感性和特异性,整体准确性始终达到或超过 90%,表明 LLM 在摘要筛选任务中具有很高的潜力。研究发现,LLM 可以在最小的人工干预下提供可靠的结果,因此是传统摘要筛选方法的经济高效替代品。

结论

虽然 LLM 工具还没有准备好完全取代人类专家进行摘要筛选,但它们在彻底改变这一过程方面显示出巨大的潜力。它们可以作为自主 AI 审查员,与人类专家的协作工作流程相结合,并整合到混合方法中,开发出提高效率的定制工具。随着技术的不断进步,LLM 有望在摘要筛选中发挥越来越重要的作用,重塑系统评价和荟萃分析研究的工作流程。

相似文献

10
Performance of a Large Language Model in Screening Citations.大语言模型在引文筛选中的表现。
JAMA Netw Open. 2024 Jul 1;7(7):e2420496. doi: 10.1001/jamanetworkopen.2024.20496.

引用本文的文献

本文引用的文献

4
Risks and Benefits of Large Language Models for the Environment.大型语言模型对环境的风险与益处
Environ Sci Technol. 2023 Mar 7;57(9):3464-3466. doi: 10.1021/acs.est.3c01106. Epub 2023 Feb 23.
5
The methodological rigour of systematic reviews in environmental health.环境健康系统综述的方法学严谨性。
Crit Rev Toxicol. 2022 Mar;52(3):167-187. doi: 10.1080/10408444.2022.2082917. Epub 2022 Jul 5.
10
Zero-Shot Learning-A Comprehensive Evaluation of the Good, the Bad and the Ugly.零样本学习:好坏丑的全面评估。
IEEE Trans Pattern Anal Mach Intell. 2019 Sep;41(9):2251-2265. doi: 10.1109/TPAMI.2018.2857768. Epub 2018 Jul 19.

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验