Curtin Medical School, Curtin University, Perth, Western Australia, Australia.
Curtin Health Innovation Research Institute, Curtin University, Perth, Western Australia, Australia.
Pharmacol Res Perspect. 2024 Feb;12(1):e1170. doi: 10.1002/prp2.1170.
Our objective was to establish and test a machine learning-based screening process that would be applicable to systematic reviews in pharmaceutical sciences. We used the SPIDER (Sample, Phenomenon of Interest, Design, Evaluation, Research type) model, a broad search strategy, and a machine learning tool (Research Screener) to identify relevant references related to y-site compatibility of 95 intravenous drugs used in neonatal intensive care settings. Two independent reviewers conducted pilot studies, including manual screening and evaluation of Research Screener, and used the kappa-coefficient for inter-reviewer reliability. After initial deduplication of the search strategy results, 27 597 references were available for screening. Research Screener excluded 1735 references, including 451 duplicate titles and 1269 reports with no abstract/title, which were manually screened. The remainder (25 862) were subject to the machine learning screening process. All eligible articles for the systematic review were extracted from <10% of the references available for screening. Moderate inter-reviewer reliability was achieved, with kappa-coefficient ≥0.75. Overall, 324 references were subject to full-text reading and 118 were deemed relevant for the systematic review. Our study showed that a broad search strategy to optimize the literature captured for systematic reviews can be efficiently screened by the semi-automated machine learning tool, Research Screener.
我们的目标是建立并测试一种基于机器学习的筛选流程,使其能够适用于制药科学领域的系统评价。我们使用了 SPIDER(样本、感兴趣现象、设计、评估、研究类型)模型、广泛的搜索策略和机器学习工具(Research Screener)来识别与新生儿重症监护环境中 95 种静脉内药物的 Y 型部位相容性相关的参考文献。两名独立的审查员进行了试点研究,包括手动筛选和 Research Screener 的评估,并使用 Kappa 系数评估了审查员之间的可靠性。在对搜索策略结果进行初步去重后,有 27597 篇参考文献可供筛选。Research Screener 排除了 1735 篇参考文献,包括 451 篇重复标题和 1269 篇没有摘要/标题的报告,这些报告需要手动筛选。其余(25862 篇)则需要进行机器学习筛选。从可供筛选的参考文献中,<10%的文献符合系统评价的纳入标准。审查员之间的可靠性达到了中等水平,Kappa 系数≥0.75。总的来说,有 324 篇参考文献需要进行全文阅读,其中 118 篇被认为与系统评价相关。我们的研究表明,广泛的搜索策略可以有效地筛选出系统评价所需的文献,而半自动的机器学习工具 Research Screener 则可以优化筛选流程。