• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

文献计量学与同行评议的一致性:来自意大利研究评估活动的证据。

On the agreement between bibliometrics and peer review: Evidence from the Italian research assessment exercises.

机构信息

Department of Economics and Statistics, University of Siena, Siena, Italy.

Department of Electrical, Computer and Biomedical Engineering, University of Pavia, Pavia, Italy.

出版信息

PLoS One. 2020 Nov 18;15(11):e0242520. doi: 10.1371/journal.pone.0242520. eCollection 2020.

DOI:10.1371/journal.pone.0242520
PMID:33206715
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7673579/
Abstract

This paper analyzes the concordance between bibliometrics and peer review. It draws evidence from the data of two experiments of the Italian governmental agency for research evaluation. The experiments were performed by the agency for validating the adoption in the Italian research assessment exercises of a dual system of evaluation, where some outputs were evaluated by bibliometrics and others by peer review. The two experiments were based on stratified random samples of journal articles. Each article was scored by bibliometrics and by peer review. The degree of concordance between the two evaluations is then computed. The correct setting of the experiments is defined by developing the design-based estimation of the Cohen's kappa coefficient and some testing procedures for assessing the homogeneity of missing proportions between strata. The results of both experiments show that for each research areas of science, technology, engineering and mathematics the degree of agreement between bibliometrics and peer review is-at most-weak at an individual article level. Thus, the outcome of the experiments does not validate the use of the dual system of evaluation in the Italian research assessments. More in general, the very weak concordance indicates that metrics should not replace peer review at the level of individual article. Hence, the use of the dual system in a research assessment might worsen the quality of information compared to the adoption of peer review only or bibliometrics only.

摘要

本文分析了文献计量学和同行评议的一致性。它从意大利政府研究评估机构的两项实验数据中得出证据。这些实验是为了验证在意大利研究评估中采用双重评估体系而进行的,其中一些产出通过文献计量学进行评估,而另一些则通过同行评议进行评估。这两项实验是基于期刊文章的分层随机抽样进行的。每篇文章都通过文献计量学和同行评议进行评分。然后计算这两种评估之间的一致性程度。通过开发基于设计的 Cohen's kappa 系数估计和一些用于评估层间缺失比例同质性的测试程序,正确设置了实验。这两项实验的结果均表明,对于科学、技术、工程和数学的各个研究领域,文献计量学和同行评议之间的一致性程度——最多——在单个文章层面上是较弱的。因此,实验结果并没有验证在意大利研究评估中使用双重评估体系的合理性。更一般地说,非常弱的一致性表明,在单个文章层面上,指标不应该取代同行评议。因此,与仅采用同行评议或文献计量学相比,在研究评估中使用双重系统可能会降低信息质量。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9aad/7673579/b3b8f9f03289/pone.0242520.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9aad/7673579/349ecf937c1a/pone.0242520.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9aad/7673579/c9fc51fb7d41/pone.0242520.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9aad/7673579/cd5717ccb690/pone.0242520.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9aad/7673579/4088802d1943/pone.0242520.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9aad/7673579/b3b8f9f03289/pone.0242520.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9aad/7673579/349ecf937c1a/pone.0242520.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9aad/7673579/c9fc51fb7d41/pone.0242520.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9aad/7673579/cd5717ccb690/pone.0242520.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9aad/7673579/4088802d1943/pone.0242520.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9aad/7673579/b3b8f9f03289/pone.0242520.g006.jpg

相似文献

1
On the agreement between bibliometrics and peer review: Evidence from the Italian research assessment exercises.文献计量学与同行评议的一致性:来自意大利研究评估活动的证据。
PLoS One. 2020 Nov 18;15(11):e0242520. doi: 10.1371/journal.pone.0242520. eCollection 2020.
2
The use and misuse of journal metrics and other citation indicators.期刊指标及其他引用指标的使用与误用。
Arch Immunol Ther Exp (Warsz). 2009 Jan-Feb;57(1):1-11. doi: 10.1007/s00005-009-0008-y. Epub 2009 Feb 14.
3
Robotic-assisted versus standard unicompartmental knee arthroplasty-evaluation of manuscript conflict of interests, funding, scientific quality and bibliometrics.机器人辅助与标准单髁膝关节置换术的比较——手稿利益冲突、资金、科学质量和文献计量学的评估。
Int Orthop. 2019 Aug;43(8):1865-1871. doi: 10.1007/s00264-018-4175-5. Epub 2018 Oct 5.
4
New developments in the use of citation analysis in research evaluation.研究评价中引用分析应用的新进展。
Arch Immunol Ther Exp (Warsz). 2009 Jan-Feb;57(1):13-8. doi: 10.1007/s00005-009-0001-5. Epub 2009 Feb 14.
5
From abstract to peer-reviewed publication: country matters.从摘要到同行评审出版物:国家因素至关重要。
Int J Cardiol. 2014 Jul 1;174(3):830-2. doi: 10.1016/j.ijcard.2014.04.172. Epub 2014 Apr 21.
6
The association between four citation metrics and peer rankings of research influence of Australian researchers in six fields of public health.四项引文计量指标与同行对澳大利亚六位公共卫生领域研究人员的研究影响力排名之间的关联。
PLoS One. 2011 Apr 6;6(4):e18521. doi: 10.1371/journal.pone.0018521.
7
[Significance of letters published in the Dutch Journal of Medicine, 1997/98].[发表于《荷兰医学杂志》1997/98年刊的信件的意义]
Ned Tijdschr Geneeskd. 2001 Mar 17;145(11):531-5.
8
A new twist on peer review.同行评议的新动向。
Elife. 2018 Jun 26;7:e36545. doi: 10.7554/eLife.36545.
9
Citations, impact factors and shady publication practices: how should the lasting clinical and social value of research really be measured?引用、影响因子与不正当发表行为:研究真正持久的临床和社会价值究竟应如何衡量?
Eur J Cancer Care (Engl). 2010 Mar;19(2):141-3. doi: 10.1111/j.1365-2354.2010.01178.x.
10
[Strenghts and weaknesses of peer review].
Med Clin (Barc). 2008 Dec;131 Suppl 5:20-4. doi: 10.1016/S0025-7753(08)76402-9.

引用本文的文献

1
Argument mining as rapid screening tool of COVID-19 literature quality: Preliminary evidence.**论点挖掘作为 COVID-19 文献质量的快速筛选工具**:初步证据。
Front Public Health. 2022 Jul 18;10:945181. doi: 10.3389/fpubh.2022.945181. eCollection 2022.

本文引用的文献

1
Kappa Coefficients for Missing Data.缺失数据的卡帕系数。
Educ Psychol Meas. 2019 Jun;79(3):558-576. doi: 10.1177/0013164418823249. Epub 2019 Jan 16.
2
Weighted kappa: nominal scale agreement with provision for scaled disagreement or partial credit.加权kappa系数:用于衡量名义尺度上的一致性,并考虑了尺度不一致或部分得分的情况。
Psychol Bull. 1968 Oct;70(4):213-20. doi: 10.1037/h0026256.
3
The measurement of observer agreement for categorical data.分类数据观察者一致性的测量。
Biometrics. 1977 Mar;33(1):159-74.