Dai Li-Yang, Jin Wen-Jie
Department of Orthopaedic Surgery, Xinhua Hospital, Shanghai Second Medical University, Shanghai, China.
Spine (Phila Pa 1976). 2005 Feb 1;30(3):354-8. doi: 10.1097/01.brs.0000152095.85927.24.
The Load Sharing Classification of spinal fractures was evaluated by 5 observers on 2 occasions.
To evaluate the interobserver and intraobserver reliability of the Load Sharing Classification of spinal fractures in the assessment of thoracolumbar burst fractures.
The Load Sharing Classification of spinal fractures provides a basis for the choice of operative approaches, but the reliability of this classification system has not been established.
The radiographic and computed tomography scan images of 45 consecutive patients with thoracolumbar burst fractures were reviewed by 5 observers on 2 different occasions 3 months apart. Interobserver reliability was assessed by comparison of the fracture classifications determined by the 5 observers. Intraobserver reliability was evaluated by comparison of the classifications determined by each observer on the first and second sessions. Ten paired interobserver and 5 intraobserver comparisons were then analyzed with use of kappa statistics.
All 5 observers agreed on the final classification for 58% and 73% of the fractures on the first and second assessments, respectively. The average kappa coefficient for the 10 paired comparisons among the 5 observers was 0.79 (range 0.73-0.89) for the first assessment and 0.84 (range 0.81-0.95) for the second assessment. Interobserver agreement improved when the 3 components of the classification system were analyzed separately, reaching an almost perfect interobserver reliability with the average kappa values of 0.90 (range 0.82-0.97) for the first assessment and 0.92 (range 0.83-1) for the second assessment. The kappa values for the 5 intraobserver comparisons ranged from 0.73 to 0.87 (average 0.78), expressing at least substantial agreement; 2 observers showed almost perfect intraobserver reliability. For the 3 components of the classification system, all observers reached almost perfect intraobserver agreement with the kappa values of 0.83 to 0.97 (average, 0.89).
Kappa statistics showed high levels of agreement when the Load Sharing Classification was used to assess thoracolumbar burst fractures. This system can be applied with excellent reliability.
5名观察者分两次对脊柱骨折的载荷分担分类进行评估。
评估在胸腰椎爆裂骨折评估中脊柱骨折载荷分担分类的观察者间及观察者内可靠性。
脊柱骨折的载荷分担分类为手术入路的选择提供了依据,但该分类系统的可靠性尚未确立。
5名观察者在相隔3个月的2个不同时间,对45例连续的胸腰椎爆裂骨折患者的X线片和计算机断层扫描图像进行回顾。通过比较5名观察者确定的骨折分类来评估观察者间可靠性。通过比较每位观察者在第一次和第二次评估时确定的分类来评估观察者内可靠性。然后使用kappa统计分析10对观察者间比较和5对观察者内比较。
在第一次和第二次评估中,所有5名观察者分别对58%和73%的骨折达成了最终分类一致。5名观察者之间10对比较的平均kappa系数,第一次评估时为0.79(范围0.73 - 0.89),第二次评估时为0.84(范围0.81 - 0.95)。当分别分析分类系统的3个组成部分时,观察者间一致性得到改善,第一次评估时平均kappa值为0.90(范围0.82 - 0.97),第二次评估时为0.92(范围0.83 - 1),达到了几乎完美的观察者间可靠性。5对观察者内比较的kappa值范围为0.73至0.87(平均0.78),表示至少有实质性一致;2名观察者表现出几乎完美的观察者内可靠性。对于分类系统的3个组成部分,所有观察者的观察者内一致性几乎完美,kappa值为0.83至0.97(平均0.89)。
当使用载荷分担分类评估胸腰椎爆裂骨折时,kappa统计显示出高度的一致性。该系统可以以极高的可靠性应用。