Adams John L, Mehrotra Ateev, Thomas J William, McGlynn Elizabeth A, Adams John L, Mehrotra Ateev, McGlynn Elizabeth A
Rand Health Q. 2012 Mar 1;2(1):3. eCollection 2012 Spring.
This article describes the methods and sensitivity analyses used by the authors in an article published in the . Purchasers are experimenting with a variety of approaches to control health care costs, including limiting network contracts to lower-cost physicians and offering patients differential copayments to encourage them to visit "high-performance" (i.e., higher-quality, lower-cost) physicians. These approaches require a method for analyzing physicians' costs and a classification system for determining which physicians have lower relative costs. There has been little analysis of the reliability of such methods. Reliability is determined by three factors: the number of observations, the variation between physicians in their use of resources, and random variation in the scores. A study of claims data from four Massachusetts health plans demonstrates that, according to the current methods of physician cost profiling, the majority of physicians did not have cost profiles that met common reliability thresholds and, importantly, reliability varied significantly by specialty. Low reliability results in a substantial chance that a given physician will be misclassified as lower-cost when he or she is not, or vice versa. Such findings raise concerns about the use of cost profiling tools and the utility of their results. It also explains the relationship between reliability measurement and misclassification for physician quality and cost measures in health care. It provides details and a practical method to calculate reliability and misclassification from the data typically available to health plans. This article builds on other RAND work on reliability and misclassification and has two main goals. First, it can serve as a tutorial for measuring reliability and misclassification. Second, it will describe the likelihood of misclassification in a situation not addressed in our prior work in which physicians are categorized using statistical testing. For any newly proposed system, the methods presented here should enable an evaluator to calculate the reliabilities and, consequently, the misclassification probabilities. It is our hope that knowing these misclassification probabilities will increase transparency about profiling methods and stimulate an informed debate about the costs and benefits of alternative profiling systems.
本文介绍了作者在发表于《 》上的一篇文章中所使用的方法和敏感性分析。购买者正在试验各种控制医疗保健成本的方法,包括将网络合同限制在低成本医生身上,并为患者提供差异化的共付额,以鼓励他们去看“高性能”(即高质量、低成本)的医生。这些方法需要一种分析医生成本的方法和一个用于确定哪些医生相对成本较低的分类系统。对于此类方法的可靠性,几乎没有进行过分析。可靠性由三个因素决定:观察次数、医生在资源使用上的差异以及分数的随机变化。一项对马萨诸塞州四个健康计划的索赔数据的研究表明,根据当前医生成本分析的方法,大多数医生的成本分析结果未达到常见的可靠性阈值,而且重要的是,可靠性因专业而异。低可靠性导致很大的可能性,即给定的医生在并非低成本时被错误分类为低成本,反之亦然。这些发现引发了对成本分析工具的使用及其结果效用的担忧。它还解释了医疗保健中医生质量和成本衡量的可靠性测量与错误分类之间的关系。它提供了从健康计划通常可获得的数据中计算可靠性和错误分类的详细信息及实用方法。本文基于兰德公司在可靠性和错误分类方面的其他工作,有两个主要目标。首先,它可以作为测量可靠性和错误分类的教程。其次,它将描述在我们之前的工作未涉及的一种情况下错误分类的可能性,即在使用统计测试对医生进行分类的情况下。对于任何新提出的系统,这里介绍的方法应使评估者能够计算可靠性,从而计算错误分类概率。我们希望了解这些错误分类概率将提高分析方法的透明度,并激发关于替代分析系统成本和效益的明智辩论。