• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过最大似然法对格韦特机遇一致性模型进行评分者间信度估计。

Interrater Reliability Estimation via Maximum Likelihood for Gwet's Chance Agreement Model.

作者信息

Westover Alek M, Westover Tara M, Westover M Brandon

机构信息

Massachusetts Institute of Technology, Boston, MA, USA.

Harvard Medical School, Beth Israel Deaconess Medical Center, Boston, MA, USA.

出版信息

Open J Stat. 2024 Oct;14(5):481-491. doi: 10.4236/ojs.2024.145021. Epub 2024 Oct 28.

DOI:10.4236/ojs.2024.145021
PMID:40342797
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC12061032/
Abstract

Interrater reliability (IRR) statistics, like Cohen's kappa, measure agreement between raters beyond what is expected by chance when classifying items into categories. While Cohen's kappa has been widely used, it has several limitations, prompting development of Gwet's agreement statistic, an alternative "kappa"statistic which models chance agreement via an "occasional guessing" model. However, we show that Gwet's formula for estimating the proportion of agreement due to chance is itself biased for intermediate levels of agreement, despite overcoming limitations of Cohen's kappa at high and low agreement levels. We derive a maximum likelihood estimator for the occasional guessing model that yields an unbiased estimator of the IRR, which we call the maximum likelihood kappa ( ). The key result is that the chance agreement probability under the occasional guessing model is simply equal to the observed rate of disagreement between raters. The statistic provides a theoretically principled approach to quantifying IRR that addresses limitations of previous coefficients. Given the widespread use of IRR measures, having an unbiased estimator is important for reliable inference across domains where rater judgments are analyzed.

摘要

评分者间信度(IRR)统计量,如科恩kappa系数,用于衡量在将项目分类时评分者之间的一致性,超出了随机预期的水平。虽然科恩kappa系数已被广泛使用,但它有几个局限性,这促使了格韦特一致性统计量的发展,这是一种替代的“kappa”统计量,它通过“偶尔猜测”模型对随机一致性进行建模。然而,我们表明,尽管格韦特在高一致性和低一致性水平上克服了科恩kappa系数的局限性,但他用于估计由于机会导致的一致性比例的公式本身在中等一致性水平上存在偏差。我们推导出了偶尔猜测模型的最大似然估计量,它产生了一个无偏的IRR估计量,我们称之为最大似然kappa( )。关键结果是,偶尔猜测模型下的随机一致性概率简单地等于评分者之间观察到的不一致率。 统计量为量化IRR提供了一种理论上有原则的方法,解决了以前 系数的局限性。鉴于IRR测量的广泛使用,拥有一个无偏估计量对于在分析评分者判断的各个领域进行可靠推断很重要。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6dbd/12061032/3fd369f13cfb/nihms-2054612-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6dbd/12061032/3fd369f13cfb/nihms-2054612-f0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6dbd/12061032/3fd369f13cfb/nihms-2054612-f0001.jpg

相似文献

1
Interrater Reliability Estimation via Maximum Likelihood for Gwet's Chance Agreement Model.通过最大似然法对格韦特机遇一致性模型进行评分者间信度估计。
Open J Stat. 2024 Oct;14(5):481-491. doi: 10.4236/ojs.2024.145021. Epub 2024 Oct 28.
2
A comparison of Cohen's Kappa and Gwet's AC1 when calculating inter-rater reliability coefficients: a study conducted with personality disorder samples.科恩氏 κ系数与格瓦特氏 AC1 系数在计算评定者间信度系数时的比较:一项对人格障碍样本进行的研究。
BMC Med Res Methodol. 2013 Apr 29;13:61. doi: 10.1186/1471-2288-13-61.
3
Gwet's AC1 is not a substitute for Cohen's kappa - A comparison of basic properties.格韦特AC1不能替代科恩kappa系数——基本特性比较
MethodsX. 2023 May 10;10:102212. doi: 10.1016/j.mex.2023.102212. eCollection 2023.
4
Interrater reliability estimators tested against true interrater reliabilities.评估者间信度估计值与真实评估者间信度进行比较。
BMC Med Res Methodol. 2022 Aug 29;22(1):232. doi: 10.1186/s12874-022-01707-5.
5
Influence of true within-herd prevalence of small ruminant lentivirus infection in goats on agreement between serological immunoenzymatic tests.山羊小反刍兽慢病毒感染的真实群体内流行率对血清学免疫酶试验之间一致性的影响
Prev Vet Med. 2017 Sep 1;144:75-80. doi: 10.1016/j.prevetmed.2017.05.017. Epub 2017 May 30.
6
Quantifying Interrater Agreement and Reliability Between Thoracic Pathologists: Paradoxical Behavior of Cohen's Kappa in the Presence of a High Prevalence of the Histopathologic Feature in Lung Cancer.量化胸科病理学家之间的评分者间一致性和可靠性:肺癌组织病理学特征高患病率情况下科恩kappa系数的矛盾行为
JTO Clin Res Rep. 2023 Dec 16;5(1):100618. doi: 10.1016/j.jtocrr.2023.100618. eCollection 2024 Jan.
7
High Agreement and High Prevalence: The Paradox of Cohen's Kappa.高一致性与高患病率:科恩kappa系数的悖论
Open Nurs J. 2017 Oct 31;11:211-218. doi: 10.2174/1874434601711010211. eCollection 2017.
8
Interrater reliability: the kappa statistic.组内一致性:kappa 统计量。
Biochem Med (Zagreb). 2012;22(3):276-82.
9
The Flexor Pollicis Longus Reflex: Interrater and Intrarater Reliability in Comparison With Established Muscle Stretch Reflexes.拇长屈肌反射:与既定肌肉牵张反射相比的评分者间和评分者内信度
Am J Phys Med Rehabil. 2021 Jun 1;100(6):539-545. doi: 10.1097/PHM.0000000000001731.
10
Interrater agreement statistics with skewed data: evaluation of alternatives to Cohen's kappa.具有偏态数据的评分者间一致性统计:对科恩kappa系数替代方法的评估
J Consult Clin Psychol. 2014 Dec;82(6):1219-27. doi: 10.1037/a0037489. Epub 2014 Aug 4.

本文引用的文献

1
Interrater agreement for Critical Care EEG Terminology.重症监护脑电学术语的评判者间一致性。
Epilepsia. 2014 Sep;55(9):1366-73. doi: 10.1111/epi.12653. Epub 2014 Jun 2.
2
A comparison of Cohen's Kappa and Gwet's AC1 when calculating inter-rater reliability coefficients: a study conducted with personality disorder samples.科恩氏 κ系数与格瓦特氏 AC1 系数在计算评定者间信度系数时的比较:一项对人格障碍样本进行的研究。
BMC Med Res Methodol. 2013 Apr 29;13:61. doi: 10.1186/1471-2288-13-61.
3
Weighted kappa: nominal scale agreement with provision for scaled disagreement or partial credit.
加权kappa系数:用于衡量名义尺度上的一致性,并考虑了尺度不一致或部分得分的情况。
Psychol Bull. 1968 Oct;70(4):213-20. doi: 10.1037/h0026256.
4
Computing inter-rater reliability and its variance in the presence of high agreement.在高度一致的情况下计算评分者间信度及其方差。
Br J Math Stat Psychol. 2008 May;61(Pt 1):29-48. doi: 10.1348/000711006X126600.
5
Understanding interobserver agreement: the kappa statistic.理解观察者间一致性:kappa统计量。
Fam Med. 2005 May;37(5):360-3.
6
Bias, prevalence and kappa.偏倚、患病率及kappa值
J Clin Epidemiol. 1993 May;46(5):423-9. doi: 10.1016/0895-4356(93)90018-v.
7
High agreement but low kappa: I. The problems of two paradoxes.高一致性但低卡帕值:I. 两个悖论的问题。
J Clin Epidemiol. 1990;43(6):543-9. doi: 10.1016/0895-4356(90)90158-l.
8
High agreement but low kappa: II. Resolving the paradoxes.高一致性但低卡帕值:II. 解决悖论
J Clin Epidemiol. 1990;43(6):551-8. doi: 10.1016/0895-4356(90)90159-m.