O'Leary Shaun, Lund Marte, Ytre-Hauge Tore Johan, Holm Sigrid Reiersen, Naess Kaja, Dalland Lars Nagelstad, McPhail Steven M
NHMRC Centre for Clinical Research Excellence in Spinal Pain, Injury and Health, University of Queensland, Brisbane, QLD 4072, Australia; Physiotherapy Department, Royal Brisbane and Women's Hospital, Queensland Health, Herston, Brisbane, QLD 4029, Australia.
NHMRC Centre for Clinical Research Excellence in Spinal Pain, Injury and Health, University of Queensland, Brisbane, QLD 4072, Australia; Norwegian Sports Medicine Clinic (NIMI), Oslo, Norway.
Physiotherapy. 2014 Mar;100(1):27-35. doi: 10.1016/j.physio.2013.08.002. Epub 2013 Nov 18.
To compare different reliability coefficients (exact agreement, and variations of the kappa (generalised, Cohen's and Prevalence Adjusted and Biased Adjusted (PABAK))) for four physiotherapists conducting visual assessments of scapulae.
Inter-therapist reliability study.
Research laboratory.
30 individuals with no history of neck or shoulder pain were recruited with no obvious significant postural abnormalities.
Ratings of scapular posture were recorded in multiple biomechanical planes under four test conditions (at rest, and while under three isometric conditions) by four physiotherapists.
The magnitude of discrepancy between the two therapist pairs was 0.04 to 0.76 for Cohen's kappa, and 0.00 to 0.86 for PABAK. In comparison, the generalised kappa provided a score between the two paired kappa coefficients. The difference between mean generalised kappa coefficients and mean Cohen's kappa (0.02) and between mean generalised kappa and PABAK (0.02) were negligible, but the magnitude of difference between the generalised kappa and paired kappa within each plane and condition was substantial; 0.02 to 0.57 for Cohen's kappa and 0.02 to 0.63 for PABAK, respectively.
Calculating coefficients for therapist pairs alone may result in inconsistent findings. In contrast, the generalised kappa provided a coefficient close to the mean of the paired kappa coefficients. These findings support an assertion that generalised kappa may lead to a better representation of reliability between three or more raters and that reliability studies only calculating agreement between two raters should be interpreted with caution. However, generalised kappa may mask more extreme cases of agreement (or disagreement) that paired comparisons may reveal.
比较四位物理治疗师对肩胛骨进行视觉评估时不同的可靠性系数(完全一致性,以及卡帕系数的变体(广义卡帕、科恩卡帕和患病率调整及偏差调整卡帕(PABAK)))。
治疗师间可靠性研究。
研究实验室。
招募了30名无颈部或肩部疼痛病史且无明显显著姿势异常的个体。
四位物理治疗师在四种测试条件下(静息状态,以及三种等长收缩条件下),在多个生物力学平面上记录肩胛骨姿势评分。
对于科恩卡帕,两组治疗师之间的差异幅度为0.04至0.76,对于PABAK为0.00至0.86。相比之下,广义卡帕给出的分数介于两组配对卡帕系数之间。广义卡帕系数均值与科恩卡帕均值之间的差异(0.02)以及广义卡帕均值与PABAK均值之间的差异(0.02)可忽略不计,但在每个平面和条件下,广义卡帕与配对卡帕之间的差异幅度较大;科恩卡帕分别为0.02至0.57,PABAK为0.02至0.63。
仅计算治疗师对之间的系数可能会导致结果不一致。相比之下,广义卡帕给出的系数接近配对卡帕系数的均值。这些发现支持这样一种观点,即广义卡帕可能会更好地反映三个或更多评估者之间的可靠性,并且仅计算两位评估者之间一致性的可靠性研究应谨慎解读。然而,广义卡帕可能会掩盖配对比较可能揭示的更极端的一致(或不一致)情况。