• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

[使用科恩kappa系数及其他系数分析分类数据的评分者间一致性]

[Analyzing interrater agreement for categorical data using Cohen's kappa and alternative coefficients].

作者信息

Wirtz M, Kutschmann M

机构信息

Institut für Psychologie, Pädagogische Hochschule Freiburg, 79117 Freiburg.

出版信息

Rehabilitation (Stuttg). 2007 Dec;46(6):370-7. doi: 10.1055/s-2007-976535.

DOI:10.1055/s-2007-976535
PMID:18188809
Abstract

Within rehabilitation research ratings are one of the most frequently used assessment procedures. For example, therapists frequently make categorical judgements aiming to get information whether specific patient characteristics prevail or not (dichotomous rating format) or which of several alternatives holds for a patient (polytomous rating format). Interrater agreement is an important prerequisite to ensure that reliable and meaningful information concerning patients' state can be inferred from the data obtained. Cohen's kappa (Cohen's kappa) is the most frequently used measure to quantify interrater agreement. The properties of Cohen's kappa are characterized and conditions for the appropriate application of kappa are clarified. Because sometimes specific properties of kappa are not appropriately considered, misleading interpretations of this measure may easily arise. This is the case because the value of Cohen's kappa is affected by information aspects that are independent of the quality of the rating process. In order to avoid such misconceptions, alternative evaluation strategies are described for dichotomous rating formats which enhance agreement analysis and thus ensure a more valid interpretation. In addition, it is shown how weighted Cohen's kappa (omega) may be used to analyze polytomous rating formats.

摘要

在康复研究中,评级是最常用的评估程序之一。例如,治疗师经常进行分类判断,旨在获取特定患者特征是否占主导地位的信息(二分评级格式),或者几种选择中的哪一种适用于某个患者(多分评级格式)。评分者间信度是确保能够从所获得的数据中推断出有关患者状况的可靠且有意义信息的重要前提。科恩卡方(Cohen's kappa)是最常用于量化评分者间信度的指标。本文阐述了科恩卡方的特性,并阐明了卡方恰当应用的条件。由于有时卡方的特定属性未得到恰当考虑,可能很容易对该指标产生误导性解释。出现这种情况是因为科恩卡方的值受到与评级过程质量无关的信息因素影响。为避免此类误解,针对二分评级格式描述了替代评估策略,这些策略可加强一致性分析,从而确保更有效的解释。此外,还展示了如何使用加权科恩卡方(omega)来分析多分评级格式。

相似文献

1
[Analyzing interrater agreement for categorical data using Cohen's kappa and alternative coefficients].[使用科恩kappa系数及其他系数分析分类数据的评分者间一致性]
Rehabilitation (Stuttg). 2007 Dec;46(6):370-7. doi: 10.1055/s-2007-976535.
2
Interobserver agreement: Cohen's kappa coefficient does not necessarily reflect the percentage of patients with congruent classifications.观察者间一致性:科恩卡方系数不一定反映分类一致的患者百分比。
Int J Clin Pharmacol Ther. 1997 Mar;35(3):93-5.
3
Clinicians are right not to like Cohen's κ.临床医生不喜欢 Cohen's κ 是对的。
BMJ. 2013 Apr 12;346:f2125. doi: 10.1136/bmj.f2125.
4
[Quality criteria of assessment scales--Cohen's kappa as measure of interrator reliability (1)].评估量表的质量标准——作为评估者信度度量的科恩kappa系数(1)
Pflege. 2004 Feb;17(1):36-46. doi: 10.1024/1012-5302.17.1.36.
5
Midwives' visual interpretation of intrapartum cardiotocographs: intra- and inter-observer agreement.助产士对产时胎心监护图的视觉解读:观察者内部及观察者间的一致性
J Adv Nurs. 2005 Oct;52(2):133-41. doi: 10.1111/j.1365-2648.2005.03575.x.
6
Reproducibility of the implant crown aesthetic index--rating aesthetics of single-implant crowns and adjacent soft tissues with regard to observer dental specialization.种植体冠美学指数的可重复性——关于观察者牙科专业对单颗种植体冠及相邻软组织美学的评级
Clin Implant Dent Relat Res. 2009 Sep;11(3):201-13. doi: 10.1111/j.1708-8208.2008.00107.x. Epub 2008 Jul 23.
7
Level of agreement between patient self-report and observer ratings of health-related quality of life communication in oncology.肿瘤学中患者自我报告与健康相关生活质量沟通的观察者评分之间的一致性水平。
Patient Educ Couns. 2007 Jan;65(1):95-100. doi: 10.1016/j.pec.2006.06.002. Epub 2006 Jul 26.
8
Low levels of interrater reliability in a standard measure of outcome in eating disorders (the modified Morgan-Russell Assessment Schedule).在进食障碍的标准结局测量方法(改良的摩根-拉塞尔评估量表)中,评分者间信度较低。
Int J Eat Disord. 1996 Jul;20(1):51-6. doi: 10.1002/(SICI)1098-108X(199607)20:1<51::AID-EAT6>3.0.CO;2-3.
9
Reliability of paramedic ratings of laryngoscopic views during endotracheal intubation.气管插管期间护理人员喉镜视野评分的可靠性
Prehosp Emerg Care. 2005 Apr-Jun;9(2):167-71. doi: 10.1080/10903120590924663.
10
Operator agreement in the use of a descriptive index of complete denture quality.
J Ir Dent Assoc. 1999;45(4):115-8.

引用本文的文献

1
Diagnoses and Treatment Recommendations-Interrater Reliability of Uroflowmetry in People with Multiple Sclerosis.诊断与治疗建议——多发性硬化症患者尿流率测定的评分者间信度
Biomedicines. 2024 Jul 18;12(7):1598. doi: 10.3390/biomedicines12071598.
2
Reliability of Two Recently Developed Procedures Assessing Biological Maturity by Ultrasound Imaging-A Pilot Study.两项近期开发的通过超声成像评估生物成熟度的程序的可靠性——一项试点研究
Children (Basel). 2024 Mar 9;11(3):326. doi: 10.3390/children11030326.
3
Inter- and intraobserver reliabilities and critical analysis of the osteoporotic fracture classification of osteoporotic vertebral body fractures.
骨质疏松性椎体骨折的骨质疏松性骨折分类的观察者间和观察者内可靠性和临界分析。
Eur Spine J. 2022 Sep;31(9):2431-2438. doi: 10.1007/s00586-022-07201-2. Epub 2022 Apr 5.
4
Self-collected and clinician-collected anal swabs show modest agreement for HPV genotyping.自我采集和临床医生采集的肛门拭子在 HPV 基因分型方面显示出适度的一致性。
PLoS One. 2021 Apr 26;16(4):e0250426. doi: 10.1371/journal.pone.0250426. eCollection 2021.
5
An instrument for quality assurance in work capacity evaluation: development, evaluation, and inter-rater reliability.一种工作能力评估质量保证工具:开发、评估和评价者间信度。
BMC Health Serv Res. 2019 Aug 9;19(1):556. doi: 10.1186/s12913-019-4387-4.
6
Prospective survey-based study on the categorization quality of hospital pharmacists' interventions using DokuPIK.基于前瞻性调查的关于使用DokuPIK对医院药师干预措施进行分类质量的研究。
Int J Clin Pharm. 2019 Apr;41(2):414-423. doi: 10.1007/s11096-019-00785-8. Epub 2019 Mar 20.
7
Development and evaluation of a standardized peer-training in the context of peer review for quality assurance in work capacity evaluation.制定和评估同行评审中标准化同行培训,以确保工作能力评估的质量。
BMC Med Educ. 2018 Jun 13;18(1):135. doi: 10.1186/s12909-018-1233-z.