Wnfried Ramirez Hillary, Chiaborelli Malena, Schönenberger Christof M, Mellor Katie, Griessbach Alexandra N, Dhiman Paula, Gandhi Pooja, Lohner Szimonetta, Agarwal Arnav, Odutayo Ayodele, Schlussel Michael M, Ravaud Philippe, Moher David, Briel Matthias, Boutron Isabelle, Hopewell Sally, Schroter Sara, Speich Benjamin
CLEAR Methods Center, Division of Clinical Epidemiology, Department Clinical Research, University Hospital Basel and University of Basel, Basel, Switzerland; Swiss Tropical and Public Health Institute (Swiss TPH), Basel, Switzerland; University of Basel, Basel, Switzerland.
CLEAR Methods Center, Division of Clinical Epidemiology, Department Clinical Research, University Hospital Basel and University of Basel, Basel, Switzerland.
J Clin Epidemiol. 2025 May 8;183:111818. doi: 10.1016/j.jclinepi.2025.111818.
Two studies randomizing manuscripts submitted to biomedical journals have previously shown that reminding peer reviewers about key reporting items did not improve the reporting quality in published articles. Within this secondary analysis of peer reviewer reports we aimed to assess at what stage the intervention failed.
We exploratively analyzed peer reviewer reports from two published randomized controlled trials (RCTs) conducted at biomedical journals. The first RCT (CONSORT-PR) assessed adherence to the Consolidated Standards of Reporting Trials (CONSORT) guideline in manuscripts presenting primary RCT results. The second RCT (SPIRIT-PR) included manuscripts presenting RCT protocols and assessed adherence to the Standard Protocol Items: Recommendations for Interventional Trials (SPIRIT) guideline. In both RCTs the control group consisted of peer reviewers receiving no reminder, whereas all reviewers in the intervention group received a reminder of the 10 most important reporting items. For this secondary analysis, we extracted from peer reviewer reports which of the ten key reporting items were mentioned by reviewers as requiring clarification. The main outcome of this secondary analysis was the difference in the mean proportion of these ten reporting items for which at least one peer reviewer requested clarification. Furthermore, we assessed how this difference changed (i) if only published manuscripts were considered and (ii) when only requested changes that were implemented by authors were considered.
We assessed peer reviewer reports from 533 manuscripts (n = 265 intervention group; n = 268 control group). Among the manuscripts in the intervention group, 21.1% (95% CI, 18.6%-23.6%) of the ten reporting items were requested for clarification, compared to 13.1% (95% CI, 18.6%-23.6%) in the control group, resulting in a mean difference of 8.0% (95% CI, 4.9%-11.1%). However, this difference diminished to 4.2% when assessing solely accepted and published manuscripts and was even further reduced to 2.6% when accounting for changes actually implemented by authors.
Reminding peer reviewers to check reporting items increased their focus on reporting guidelines, leading to more reporting-related requests in their reviews. However, the effect was strongly diluted during the peer review process due to rejected articles and requests not implemented by authors.
When new research is submitted to a journal, other experts in the field (peer reviewers) check the research to make sure it's reliable and clear. Among others, one important part of this process is ensuring that researchers follow reporting guidelines about what information should be included in their papers so that the readers can understand how the research was conducted. We wanted to find out if reminding peer reviewers to focus on the key parts of these guidelines (ie, 10 most important items) would help to improve the reporting quality of published research papers. For this purpose, we conducted two studies in which we randomized manuscripts to either an intervention group or a control group. In the intervention group, the peer reviewers from half of the included manuscript received such a reminder (ie, asking them to check whether the 10 most important reporting items are well described in the manuscript), whereas peer reviewers in the control group did not receive a reminder. Within our previously published main results of these studies we saw that the reporting quality of the published articles did not improve with this intervention. To find out why this approach did not work, we looked closer at the individual reports from peer reviewers and checked how often reviewers asked for these important details and whether authors made the necessary changes. We found that reminders did lead to more requests about reporting items from peer reviewers. However, as a high proportion of peer-reviewed articles is rejected during the peer review process and because not all requests for improvements are addressed by authors, this effect was not visible anymore (ie, "diluted") when assessing published research articles.
此前两项对提交给生物医学期刊的稿件进行随机分组的研究表明,提醒同行评审员关注关键报告项目并不能提高已发表文章的报告质量。在对同行评审报告的这项二次分析中,我们旨在评估干预在哪个阶段失败。
我们对生物医学期刊发表的两项随机对照试验(RCT)的同行评审报告进行了探索性分析。第一项RCT(CONSORT-PR)评估了呈现RCT主要结果的稿件对《报告试验的统一标准》(CONSORT)指南的遵循情况。第二项RCT(SPIRIT-PR)纳入了呈现RCT方案的稿件,并评估了对《干预性试验标准方案项目:建议》(SPIRIT)指南的遵循情况。在这两项RCT中,对照组由未收到提醒的同行评审员组成,而干预组的所有评审员都收到了10项最重要报告项目的提醒。对于这项二次分析,我们从同行评审报告中提取评审员提到需要澄清的10项关键报告项目中的哪些项目。这项二次分析的主要结果是这10项报告项目中至少有一位同行评审员要求澄清的项目的平均比例差异。此外,我们评估了这种差异如何变化:(i)如果只考虑已发表的稿件;(ii)当只考虑作者实施的所要求的修改时。
我们评估了533篇稿件的同行评审报告(干预组n = 265;对照组n = 268)。在干预组的稿件中,10项报告项目中有21.1%(95%CI,18.6%-23.6%)被要求澄清,而对照组为13.1%(95%CI,18.6%-23.6%),平均差异为8.0%(95%CI,4.9%-11.1%)。然而,在仅评估被接受并发表的稿件时,这种差异降至4.2%,而在考虑作者实际实施的修改时,差异甚至进一步降至2.6%。
提醒同行评审员检查报告项目增加了他们对报告指南的关注,导致他们在评审中提出更多与报告相关的要求。然而,由于被拒稿件以及作者未实施的要求,在同行评审过程中这种效果被大大削弱。
当新的研究提交给期刊时,该领域的其他专家(同行评审员)会检查该研究以确保其可靠且清晰。在这个过程中,一个重要部分是确保研究人员遵循关于论文应包含哪些信息的报告指南,以便读者能够理解研究是如何进行 的。我们想弄清楚提醒同行评审员关注这些指南的关键部分(即10项最重要的项目)是否有助于提高已发表研究论文 的报告质量。为此,我们进行了两项研究,将稿件随机分为干预组或对照组。在干预组中,所纳入稿件的一半的同行评审员收到了这样的提醒(即要求他们检查稿件中是否很好地描述了10项最重要的报告项目),而对照组的同行评审员没有收到提醒。在我们之前发表的这些研究的主要结果中,我们发现通过这种干预,已发表文章的报告质量并未提高。为了弄清楚这种方法为何不起作用,我们更仔细地查看了同行评审员的个人报告,并检查评审员多久会要求提供这些重要细节以及作者是否做出了必要的修改。我们发现提醒确实导致同行评审员提出了更多关于报告项目的要求。然而,由于在同行评审过程中有很大比例的同行评审文章被拒,并且并非所有改进要求都得到作者的回应,所以在评估已发表的研究文章时,这种效果不再明显(即“被削弱”)。