• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

CONSORT-TM:用于评估随机对照试验出版物完整性的文本分类模型。

CONSORT-TM: Text classification models for assessing the completeness of randomized controlled trial publications.

作者信息

Jiang Lan, Lan Mengfei, Menke Joe D, Vorland Colby J, Kilicoglu Halil

机构信息

School of Information Sciences, University of Illinois Urbana-Champaign, Champaign, IL, USA.

Indiana University, School of Public Health, Bloomington, IN, USA.

出版信息

medRxiv. 2024 Apr 1:2024.03.31.24305138. doi: 10.1101/2024.03.31.24305138.

DOI:10.1101/2024.03.31.24305138
PMID:38633775
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11023672/
Abstract

OBJECTIVE

To develop text classification models for determining whether the checklist items in the CONSORT reporting guidelines are reported in randomized controlled trial publications.

MATERIALS AND METHODS

Using a corpus annotated at the sentence level with 37 fine-grained CONSORT items, we trained several sentence classification models (PubMedBERT fine-tuning, BioGPT fine-tuning, and in-context learning with GPT-4) and compared their performance. To address the problem of small training dataset, we used several data augmentation methods (EDA, UMLS-EDA, text generation and rephrasing with GPT-4) and assessed their impact on the fine-tuned PubMedBERT model. We also fine-tuned PubMedBERT models limited to checklist items associated with specific sections (e.g., Methods) to evaluate whether such models could improve performance compared to the single full model. We performed 5-fold cross-validation and report precision, recall, F score, and area under curve (AUC).

RESULTS

Fine-tuned PubMedBERT model that takes as input the sentence and the surrounding sentence representations and uses section headers yielded the best overall performance (0.71 micro-F, 0.64 macro-F). Data augmentation had limited positive effect, UMLS-EDA yielding slightly better results than data augmentation using GPT-4. BioGPT fine-tuning and GPT-4 in-context learning exhibited suboptimal results. Methods-specific model yielded higher performance for methodology items, other section-specific models did not have significant impact.

CONCLUSION

Most CONSORT checklist items can be recognized reasonably well with the fine-tuned PubMedBERT model but there is room for improvement. Improved models can underpin the journal editorial workflows and CONSORT adherence checks and can help authors in improving the reporting quality and completeness of their manuscripts.

摘要

目的

开发文本分类模型,以确定随机对照试验出版物中是否报告了CONSORT报告指南中的清单项目。

材料与方法

使用一个在句子级别标注了37个细粒度CONSORT项目的语料库,我们训练了几个句子分类模型(PubMedBERT微调、BioGPT微调以及使用GPT-4的上下文学习)并比较了它们的性能。为了解决训练数据集较小的问题,我们使用了几种数据增强方法(EDA、UMLS-EDA、文本生成以及使用GPT-4进行改写)并评估了它们对微调后的PubMedBERT模型的影响。我们还对仅限于与特定部分(如方法部分)相关的清单项目的PubMedBERT模型进行了微调,以评估与单一完整模型相比,此类模型是否能提高性能。我们进行了5折交叉验证,并报告了精确率、召回率、F分数和曲线下面积(AUC)。

结果

以句子及其周围句子表示作为输入并使用章节标题的微调后的PubMedBERT模型产生了最佳的总体性能(微F值为0.71,宏F值为0.64)。数据增强的积极效果有限,UMLS-EDA产生的结果略优于使用GPT-4的数据增强。BioGPT微调以及GPT-4上下文学习表现出次优结果。特定于方法部分的模型在方法学项目上产生了更高的性能,其他特定于部分的模型没有显著影响。

结论

经过微调的PubMedBERT模型能够较好地识别大多数CONSORT清单项目,但仍有改进空间。改进后的模型可以支持期刊编辑工作流程和CONSORT依从性检查,并有助于作者提高其稿件的报告质量和完整性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fb32/11023672/0907a741935b/nihpp-2024.03.31.24305138v1-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fb32/11023672/0907a741935b/nihpp-2024.03.31.24305138v1-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/fb32/11023672/0907a741935b/nihpp-2024.03.31.24305138v1-f0001.jpg

相似文献

1
CONSORT-TM: Text classification models for assessing the completeness of randomized controlled trial publications.CONSORT-TM:用于评估随机对照试验出版物完整性的文本分类模型。
medRxiv. 2024 Apr 1:2024.03.31.24305138. doi: 10.1101/2024.03.31.24305138.
2
Text classification models for assessing the completeness of randomized controlled trial publications based on CONSORT reporting guidelines.基于 CONSORT 报告规范的评估随机对照试验出版物完整性的文本分类模型。
Sci Rep. 2024 Sep 17;14(1):21721. doi: 10.1038/s41598-024-72130-7.
3
Toward assessing clinical trial publications for reporting transparency.迈向评估临床试验出版物报告的透明度。
J Biomed Inform. 2021 Apr;116:103717. doi: 10.1016/j.jbi.2021.103717. Epub 2021 Feb 26.
4
Automatic categorization of self-acknowledged limitations in randomized controlled trial publications.自我承认的随机对照试验出版物局限性的自动分类。
J Biomed Inform. 2024 Apr;152:104628. doi: 10.1016/j.jbi.2024.104628. Epub 2024 Mar 26.
5
Consolidated standards of reporting trials (CONSORT) and the completeness of reporting of randomised controlled trials (RCTs) published in medical journals.试验报告的统一标准(CONSORT)以及医学期刊上发表的随机对照试验(RCT)的报告完整性。
Cochrane Database Syst Rev. 2012 Nov 14;11(11):MR000030. doi: 10.1002/14651858.MR000030.pub2.
6
Reporting Quality of Randomized Controlled Trials of Periodontal Diseases in Journal Abstracts-A Cross-sectional Survey and Bibliometric Analysis.期刊摘要中牙周病随机对照试验的报告质量:横断面调查和文献计量分析。
J Evid Based Dent Pract. 2018 Jun;18(2):130-141.e22. doi: 10.1016/j.jebdp.2017.08.005. Epub 2017 Sep 21.
7
Reminding Peer Reviewers of Reporting Guideline Items to Improve Completeness in Published Articles: Primary Results of 2 Randomized Trials.提醒同行评审员注意报告指南条目,以提高已发表文章的完整性:两项随机试验的主要结果。
JAMA Netw Open. 2023 Jun 1;6(6):e2317651. doi: 10.1001/jamanetworkopen.2023.17651.
8
Leveraging pre-trained language models for mining microbiome-disease relationships.利用预训练语言模型挖掘微生物组-疾病关系。
BMC Bioinformatics. 2023 Jul 19;24(1):290. doi: 10.1186/s12859-023-05411-z.
9
Investigating the impact of weakly supervised data on text mining models of publication transparency: a case study on randomized controlled trials.研究弱监督数据对出版物透明度文本挖掘模型的影响:以随机对照试验为例。
AMIA Jt Summits Transl Sci Proc. 2022 May 23;2022:254-263. eCollection 2022.
10
Impact of a web-based tool (WebCONSORT) to improve the reporting of randomised trials: results of a randomised controlled trial.基于网络的工具(WebCONSORT)对改善随机试验报告的影响:一项随机对照试验的结果
BMC Med. 2016 Nov 28;14(1):199. doi: 10.1186/s12916-016-0736-x.

本文引用的文献

1
Opportunities and challenges for ChatGPT and large language models in biomedicine and health.ChatGPT 和大型语言模型在生物医学和健康领域的机遇与挑战。
Brief Bioinform. 2023 Nov 22;25(1). doi: 10.1093/bib/bbad493.
2
Methodology reporting improved over time in 176,469 randomized controlled trials.方法学报告在 176469 项随机对照试验中随着时间的推移而改善。
J Clin Epidemiol. 2023 Oct;162:19-28. doi: 10.1016/j.jclinepi.2023.08.004. Epub 2023 Aug 9.
3
Putting ChatGPT's Medical Advice to the (Turing) Test: Survey Study.对ChatGPT的医学建议进行(图灵)测试:调查研究。
JMIR Med Educ. 2023 Jul 10;9:e46939. doi: 10.2196/46939.
4
Methodological information extraction from randomized controlled trial publications: a pilot study.从随机对照试验出版物中提取方法学信息:一项初步研究。
AMIA Annu Symp Proc. 2023 Apr 29;2022:542-551. eCollection 2022.
5
BioGPT: generative pre-trained transformer for biomedical text generation and mining.BioGPT:用于生物医学文本生成和挖掘的生成式预训练转换器。
Brief Bioinform. 2022 Nov 19;23(6). doi: 10.1093/bib/bbac409.
6
Investigating the impact of weakly supervised data on text mining models of publication transparency: a case study on randomized controlled trials.研究弱监督数据对出版物透明度文本挖掘模型的影响:以随机对照试验为例。
AMIA Jt Summits Transl Sci Proc. 2022 May 23;2022:254-263. eCollection 2022.
7
Is the future of peer review automated?同行评审的未来是自动化的吗?
BMC Res Notes. 2022 Jun 11;15(1):203. doi: 10.1186/s13104-022-06080-6.
8
The PRISMA 2020 statement: an updated guideline for reporting systematic reviews.PRISMA 2020 声明:系统评价报告的更新指南。
BMJ. 2021 Mar 29;372:n71. doi: 10.1136/bmj.n71.
9
Toward assessing clinical trial publications for reporting transparency.迈向评估临床试验出版物报告的透明度。
J Biomed Inform. 2021 Apr;116:103717. doi: 10.1016/j.jbi.2021.103717. Epub 2021 Feb 26.
10
Automated screening of COVID-19 preprints: can we help authors to improve transparency and reproducibility?新冠疫情预印本的自动化筛选:我们能否帮助作者提高透明度和可重复性?
Nat Med. 2021 Jan;27(1):6-7. doi: 10.1038/s41591-020-01203-7.