Seigel D G, Podgor M J, Remaley N A
Biometry and Epidemiology Program, National Eye Institute, National Institutes of Health, Bethesda, MD 20892.
Am J Epidemiol. 1992 Mar 1;135(5):571-8. doi: 10.1093/oxfordjournals.aje.a116324.
A model was developed for a simple clinical trial in which graders had defined probabilities of misclassifying pathologic material to disease present or absent. The authors compared Kappa between graders, and efficiency and bias in the clinical trial in the presence of misclassification. Though related to bias and efficiency, Kappa did not predict these two statistics well. These results pertain generally to evaluation of systems for encoding medical information, and the relevance of Kappa in determining whether such systems are ready for use in comparative studies. The authors conclude that, by itself, Kappa is not informative enough to evaluate the appropriateness of a grading scheme for comparative studies. Additional, and perhaps difficult, questions must be addressed for such evaluation.
针对一项简单的临床试验开发了一个模型,在该试验中,分级人员对将病理材料误分类为疾病存在或不存在有明确的概率。作者比较了分级人员之间的kappa值,以及在存在错误分类的情况下临床试验中的效率和偏差。尽管kappa值与偏差和效率相关,但它并不能很好地预测这两个统计数据。这些结果普遍适用于对医学信息编码系统的评估,以及kappa值在确定此类系统是否准备好在比较研究中使用方面的相关性。作者得出结论,仅凭kappa值本身,不足以评估用于比较研究的分级方案的适用性。对于此类评估,必须解决其他或许困难的问题。