• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

再探休伯尔氏多评估者 κ 系数。

Hubert's multi-rater kappa revisited.

机构信息

Bioestadística, Facultad de Medicina, Universidad de Granada, Granada, Spain.

Centro Universitario de la Defensa - ENM, Universidad de Vigo, Vigo, Pontevedra, Spain.

出版信息

Br J Math Stat Psychol. 2020 Feb;73(1):1-22. doi: 10.1111/bmsp.12167. Epub 2019 May 6.

DOI:10.1111/bmsp.12167
PMID:31056757
Abstract

There is a frequent need to measure the degree of agreement among R observers who independently classify n subjects within K nominal or ordinal categories. The most popular methods are usually kappa-type measurements. When R = 2, Cohen's kappa coefficient (weighted or not) is well known. When defined in the ordinal case while assuming quadratic weights, Cohen's kappa has the advantage of coinciding with the intraclass and concordance correlation coefficients. When R > 2, there are more discrepancies because the definition of the kappa coefficient depends on how the phrase 'an agreement has occurred' is interpreted. In this paper, Hubert's interpretation, that 'an agreement occurs if and only if all raters agree on the categorization of an object', is used, which leads to Hubert's (nominal) and Schuster and Smith's (ordinal) kappa coefficients. Formulae for the large-sample variances for the estimators of all these coefficients are given, allowing the latter to illustrate the different ways of carrying out inference and, with the use of simulation, to select the optimal procedure. In addition, it is shown that Schuster and Smith's kappa coefficient coincides with the intraclass and concordance correlation coefficients if the first coefficient is also defined assuming quadratic weights.

摘要

在医学研究中,经常需要衡量 R 位观察者在 K 个名义或有序类别中独立分类 n 个对象的一致性程度。最常用的方法通常是 Kappa 型测量。当 R=2 时,Cohen 的 Kappa 系数(加权或未加权)是众所周知的。当在有序情况下定义并假设二次权重时,Cohen 的 Kappa 系数与组内相关系数和一致性相关系数一致。当 R>2 时,会有更多的差异,因为 Kappa 系数的定义取决于如何解释“已经达成一致”这句话。在本文中,使用了 Hubert 的解释,即“如果所有评分者都同意对一个对象的分类,则达成一致”,这导致了 Hubert(名义)和 Schuster 和 Smith(有序)的 Kappa 系数。给出了所有这些系数估计量的大样本方差公式,允许后者说明进行推理的不同方法,并通过模拟选择最佳程序。此外,如果也假设二次权重来定义第一个系数,则 Schuster 和 Smith 的 Kappa 系数与组内相关系数和一致性相关系数一致。

相似文献

1
Hubert's multi-rater kappa revisited.再探休伯尔氏多评估者 κ 系数。
Br J Math Stat Psychol. 2020 Feb;73(1):1-22. doi: 10.1111/bmsp.12167. Epub 2019 May 6.
2
Measures of Agreement with Multiple Raters: Fréchet Variances and Inference.多评分者一致性的度量:Fréchet 方差和推断。
Psychometrika. 2024 Jun;89(2):517-541. doi: 10.1007/s11336-023-09945-2. Epub 2024 Jan 8.
3
An Alternative Interpretation of the Linearly Weighted Kappa Coefficients for Ordinal Data.有序数据线性加权kappa系数的另一种解释。
Psychometrika. 2018 May 15. doi: 10.1007/s11336-018-9621-1.
4
Dependence of weighted kappa coefficients on the number of categories.加权kappa系数对类别数量的依赖性。
Epidemiology. 1996 Mar;7(2):199-202. doi: 10.1097/00001648-199603000-00016.
5
Robustness of -type coefficients for clinical agreement.- 型系数用于临床一致性的稳健性。
Stat Med. 2022 May 20;41(11):1986-2004. doi: 10.1002/sim.9341. Epub 2022 Feb 6.
6
A comparison of Cohen's Kappa and Gwet's AC1 when calculating inter-rater reliability coefficients: a study conducted with personality disorder samples.科恩氏 κ系数与格瓦特氏 AC1 系数在计算评定者间信度系数时的比较:一项对人格障碍样本进行的研究。
BMC Med Res Methodol. 2013 Apr 29;13:61. doi: 10.1186/1471-2288-13-61.
7
Weighted least-squares approach for comparing correlated kappa.用于比较相关kappa的加权最小二乘法。
Biometrics. 2002 Dec;58(4):1012-9. doi: 10.1111/j.0006-341x.2002.01012.x.
8
Assessing the inter-rater agreement for ordinal data through weighted indexes.通过加权指数评估有序数据的评分者间一致性。
Stat Methods Med Res. 2016 Dec;25(6):2611-2633. doi: 10.1177/0962280214529560. Epub 2014 Apr 16.
9
Testing the Difference of Correlated Agreement Coefficients for Statistical Significance.检验相关一致性系数差异的统计学显著性。
Educ Psychol Meas. 2016 Aug;76(4):609-637. doi: 10.1177/0013164415596420. Epub 2015 Jul 28.
10
The impact of grey zones on the accuracy of agreement measures for ordinal tables.灰区对有序表一致性度量准确性的影响。
BMC Med Res Methodol. 2021 Apr 14;21(1):70. doi: 10.1186/s12874-021-01248-3.

引用本文的文献

1
Impact of first ambulation time on unilateral biportal endoscopy in lumbar disc herniation: a systematic review and meta-analysis.首次下床活动时间对腰椎间盘突出症单侧双孔通道内镜手术的影响:一项系统评价与Meta分析
Int J Surg. 2025 Sep 1;111(9):6362-6373. doi: 10.1097/JS9.0000000000002686. Epub 2025 Jun 20.
2
Performance of the Large Language Models in African rheumatology: a diagnostic test accuracy study of ChatGPT-4, Gemini, Copilot, and Claude artificial intelligence.大语言模型在非洲风湿病学中的表现:ChatGPT-4、Gemini、Copilot和Claude人工智能的诊断测试准确性研究
BMC Rheumatol. 2025 May 16;9(1):54. doi: 10.1186/s41927-025-00512-z.
3
Some common statistical methods for assessing rater agreement in radiological studies.
放射学研究中评估评分者一致性的一些常用统计方法。
Acta Radiol. 2025 Jun;66(6):675-683. doi: 10.1177/02841851251319666. Epub 2025 Feb 23.
4
Measures of Agreement with Multiple Raters: Fréchet Variances and Inference.多评分者一致性的度量:Fréchet 方差和推断。
Psychometrika. 2024 Jun;89(2):517-541. doi: 10.1007/s11336-023-09945-2. Epub 2024 Jan 8.
5
Inter-Rater Agreement in Assessing Risk of Bias in Melanoma Prediction Studies Using the Prediction Model Risk of Bias Assessment Tool (PROBAST): Results from a Controlled Experiment on the Effect of Specific Rater Training.使用预测模型偏倚风险评估工具(PROBAST)评估黑色素瘤预测研究中偏倚风险的评分者间一致性:关于特定评分者培训效果的对照实验结果
J Clin Med. 2023 Mar 2;12(5):1976. doi: 10.3390/jcm12051976.