• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

相似文献

1
Computing Inter-Rater Reliability for Observational Data: An Overview and Tutorial.计算观测数据的评分者间信度:概述与教程
Tutor Quant Methods Psychol. 2012;8(1):23-34. doi: 10.20982/tqmp.08.1.p023.
2
Kappa statistic considerations in evaluating inter-rater reliability between two raters: which, when and context matters.评价两位评分者之间的评分者间可靠性时的 Kappa 统计量考虑因素:何时、何种情况下以及上下文很重要。
BMC Cancer. 2023 Aug 25;23(1):799. doi: 10.1186/s12885-023-11325-z.
3
Evaluating inter-rater reliability in the context of "Sysmex UN2000 detection of protein/creatinine ratio and of renal tubular epithelial cells can be used for screening lupus nephritis": a statistical examination.评估“Sysmex UN2000 检测蛋白/肌酐比值和肾小管上皮细胞可用于狼疮肾炎筛查”中的组内信度:一项统计学检验。
BMC Nephrol. 2024 Mar 13;25(1):94. doi: 10.1186/s12882-024-03540-y.
4
Inter- and intra-rater reliability for measurement of range of motion in joints included in three hypermobility assessment methods.三种关节过度活动评估方法中所包含关节活动范围测量的评分者间信度和评分者内信度。
BMC Musculoskelet Disord. 2018 Oct 17;19(1):376. doi: 10.1186/s12891-018-2290-5.
5
A reliability-generalization study of journal peer reviews: a multilevel meta-analysis of inter-rater reliability and its determinants.期刊同行评审的可靠性综合研究:评分者间可靠性及其决定因素的多级元分析。
PLoS One. 2010 Dec 14;5(12):e14331. doi: 10.1371/journal.pone.0014331.
6
Reliability of Observational Assessment Methods for Outcome-based Assessment of Surgical Skill: Systematic Review and Meta-analyses.基于观察的手术技能结果评估方法的可靠性:系统评价和荟萃分析。
J Surg Educ. 2020 Jan-Feb;77(1):189-201. doi: 10.1016/j.jsurg.2019.07.007. Epub 2019 Aug 20.
7
Inter-rater reliability of AMSTAR is dependent on the pair of reviewers.AMSTAR的评分者间信度取决于评审者对。
BMC Med Res Methodol. 2017 Jul 11;17(1):98. doi: 10.1186/s12874-017-0380-y.
8
Use of hand-held Doppler ultrasound examination by podiatrists: a reliability study.足病医生使用手持式多普勒超声检查:一项可靠性研究。
J Foot Ankle Res. 2015 Aug 12;8:36. doi: 10.1186/s13047-015-0097-2. eCollection 2015.
9
Midwives' visual interpretation of intrapartum cardiotocographs: intra- and inter-observer agreement.助产士对产时胎心监护图的视觉解读:观察者内部及观察者间的一致性
J Adv Nurs. 2005 Oct;52(2):133-41. doi: 10.1111/j.1365-2648.2005.03575.x.
10
Virtual audits of the urban streetscape: comparing the inter-rater reliability of GigaPan® to Google Street View.虚拟城市街景审核:比较 GigaPan® 与谷歌街景的评分者间信度。
Int J Health Geogr. 2020 Aug 12;19(1):31. doi: 10.1186/s12942-020-00226-0.

引用本文的文献

1
The Efficacy of Acceptance and Commitment Therapy for Transitional-Age Youth: A Meta-analysis.接纳与承诺疗法对过渡年龄青少年的疗效:一项元分析。
Clin Child Fam Psychol Rev. 2025 Sep 3. doi: 10.1007/s10567-025-00543-5.
2
Intra- and inter-observer reliability and repeatability of the metatarsus adductus angle in childhood: A concordance study.儿童期内收足跖骨角的观察者间及观察者内可靠性与可重复性:一项一致性研究。
Pediatr Radiol. 2025 Aug 25. doi: 10.1007/s00247-025-06375-3.
3
The Youth-Focused Cognitive and Behavioral Therapy Adherence Measure: A Multi-Informant, Transdiagnostic, Trans-Manual, Adherence Monitoring Tool.以青少年为重点的认知行为疗法依从性测量:一种多信息源、跨诊断、跨手册的依从性监测工具。
Adm Policy Ment Health. 2025 Aug 14. doi: 10.1007/s10488-025-01462-x.
4
"A Dog Brings Benefits No Matter Where It's from": UK Residents' Understanding of the Benefits and Risks of Importing Puppies from Romania to the UK.“无论来自何处,狗狗都能带来益处”:英国居民对从罗马尼亚进口幼犬至英国的益处与风险的理解
Animals (Basel). 2025 Jul 25;15(15):2192. doi: 10.3390/ani15152192.
5
Lumbosacral spinal cord functional connectivity at rest: From feasibility to reliability.静息状态下腰骶脊髓的功能连接:从可行性到可靠性
Imaging Neurosci (Camb). 2024 Sep 5;2. doi: 10.1162/imag_a_00286. eCollection 2024.
6
Physiotherapists' reliability of inter-recti distance measurement with real-time ultrasound across a mixed women population sample.物理治疗师通过实时超声对混合女性人群样本进行腹直肌间距测量的可靠性。
Womens Health (Lond). 2025 Jan-Dec;21:17455057251361999. doi: 10.1177/17455057251361999. Epub 2025 Aug 12.
7
Development and computer-assisted validation of a radio frequency identification system for tracking individual chicken visits to functional areas.用于跟踪单个鸡进入功能区域的射频识别系统的开发与计算机辅助验证。
Poult Sci. 2025 Jul 31;104(10):105627. doi: 10.1016/j.psj.2025.105627.
8
Spotlighting healthcare frontline workers´ perceptions on artificial intelligence across the globe.聚焦全球医疗一线工作者对人工智能的看法。
Npj Health Syst. 2025;2(1):28. doi: 10.1038/s44401-025-00034-3. Epub 2025 Jul 30.
9
A mixed-methods framework for assessing differentiated instruction implementation barriers in EFL secondary education contexts.一个用于评估外语非母语(EFL)中等教育背景下差异化教学实施障碍的混合方法框架。
MethodsX. 2025 Jun 21;15:103457. doi: 10.1016/j.mex.2025.103457. eCollection 2025 Dec.
10
A comparative study of student performance in all-ceramic crown preparation by clinical-phase students.临床阶段学生全瓷冠修复预备操作中学生表现的比较研究。
BMC Med Educ. 2025 Jul 15;25(1):1057. doi: 10.1186/s12909-025-07512-0.

本文引用的文献

1
Weighted kappa: nominal scale agreement with provision for scaled disagreement or partial credit.加权kappa系数:用于衡量名义尺度上的一致性,并考虑了尺度不一致或部分得分的情况。
Psychol Bull. 1968 Oct;70(4):213-20. doi: 10.1037/h0026256.
2
Intraclass correlations: uses in assessing rater reliability.组内相关系数:在评估评分者可靠性中的应用。
Psychol Bull. 1979 Mar;86(2):420-8. doi: 10.1037//0033-2909.86.2.420.
3
Ill-structured measurement designs in organizational research: implications for estimating interrater reliability.组织研究中结构不良的测量设计:对评估评分者间信度的影响。
J Appl Psychol. 2008 Sep;93(5):959-81. doi: 10.1037/0021-9010.93.5.959.
4
Bias, prevalence and kappa.偏倚、患病率及kappa值
J Clin Epidemiol. 1993 May;46(5):423-9. doi: 10.1016/0895-4356(93)90018-v.
5
The kappa coefficient of agreement for multiple observers when the number of subjects is small.当受试者数量较少时,多位观察者之间的一致性kappa系数。
Biometrics. 1986 Dec;42(4):883-93.
6
The measurement of observer agreement for categorical data.分类数据观察者一致性的测量。
Biometrics. 1977 Mar;33(1):159-74.

计算观测数据的评分者间信度:概述与教程

Computing Inter-Rater Reliability for Observational Data: An Overview and Tutorial.

作者信息

Hallgren Kevin A

机构信息

University of New Mexico, Department of Psychology.

出版信息

Tutor Quant Methods Psychol. 2012;8(1):23-34. doi: 10.20982/tqmp.08.1.p023.

DOI:10.20982/tqmp.08.1.p023
PMID:22833776
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC3402032/
Abstract

Many research designs require the assessment of inter-rater reliability (IRR) to demonstrate consistency among observational ratings provided by multiple coders. However, many studies use incorrect statistical procedures, fail to fully report the information necessary to interpret their results, or do not address how IRR affects the power of their subsequent analyses for hypothesis testing. This paper provides an overview of methodological issues related to the assessment of IRR with a focus on study design, selection of appropriate statistics, and the computation, interpretation, and reporting of some commonly-used IRR statistics. Computational examples include SPSS and R syntax for computing Cohen's kappa and intra-class correlations to assess IRR.

摘要

许多研究设计需要评估评分者间信度(IRR),以证明多个编码员提供的观察评分之间的一致性。然而,许多研究使用了不正确的统计程序,未能充分报告解释其结果所需的信息,或者没有说明IRR如何影响其后续假设检验分析的功效。本文概述了与IRR评估相关的方法学问题,重点关注研究设计、适当统计方法的选择,以及一些常用IRR统计量的计算、解释和报告。计算示例包括用于计算科恩kappa系数和组内相关性以评估IRR的SPSS和R语法。