Rifkin William D, Rifkin Arthur
Deaprtment of Medicine, Yale Primary Care Residency Program, Waterbury Hospital, 64 Robbins Street, Waterbury, CT 06721, USA.
Mt Sinai J Med. 2005 Jan;72(1):47-9.
There is interest in the use of "standardized patients" to assist in evaluating medical trainees' clinical skills, which may be difficult to evaluate with written exams alone. Previous studies of the validity of observed structured clinical exams have found low correlation with various written exams as well as with faculty evaluations. Since the United States Medical Licensing Examination (USMLE) results are often used by training programs in the selection of applicants, we assessed the correlation between performance on an observed structured clinical exam and the USMLE, steps 1 and 2, for internal medicine housestaff.
We collected scores on the USMLE, steps 1 and 2, and the overall score from a required standardized patient encounter for all PGY-1 trainees, in a single urban teaching hospital. Pearson coefficients were used to compare the USMLE and observed structured clinical exam performance.
The two steps of the USMLE correlated with each other to a large extent (r=0.65, df=30, p=0.0001). However, both steps of the USMLE correlated poorly with the observed structured clinical exam (step 1 r=0.2, df=32, p=0.27; step 2 r=0.09, df=30, p=0.61).
The low correlation between the USMLE and performance on a structured clinical exam suggests that either the written exam is a poor predictor of actual clinical performance, the small window of clinical skills measured by the structured clinical exam is inadequate, or the two methods evaluate different skill sets entirely. Our findings are consistent with previous work finding low correlations between structured clinical exams and accepted common means of evaluation, such as faculty evaluations, other written exams and program director assessments. The medical education community needs to develop an objective, valid method of measuring important, yet subjective, skill-sets such as interpersonal communication, empathy and efficient data collection.
人们对使用“标准化病人”来辅助评估医学实习生的临床技能很感兴趣,因为仅通过笔试可能难以评估这些技能。先前关于观察性结构化临床考试有效性的研究发现,其与各种笔试以及教师评估的相关性较低。由于培训项目在选拔申请者时经常使用美国医师执照考试(USMLE)的结果,我们评估了内科住院医师在观察性结构化临床考试中的表现与USMLE第一步和第二步之间的相关性。
我们收集了一家城市教学医院所有PGY - 1实习生的USMLE第一步和第二步的成绩,以及一次必需的标准化病人会诊的总成绩。使用Pearson系数来比较USMLE成绩和观察性结构化临床考试的表现。
USMLE的两个步骤在很大程度上相互关联(r = 0.65,自由度 = 30,p = 0.0001)。然而,USMLE的两个步骤与观察性结构化临床考试的相关性都很差(第一步r = 0.2,自由度 = 32,p = 0.27;第二步r = 0.09,自由度 = 30,p = 0.61)。
USMLE与结构化临床考试表现之间的低相关性表明,要么笔试不能很好地预测实际临床能力,要么结构化临床考试所测量的临床技能范围过小,要么这两种方法评估的是完全不同的技能集。我们的研究结果与之前的工作一致,即结构化临床考试与公认的常见评估方式(如教师评估、其他笔试和项目主任评估)之间的相关性较低。医学教育界需要开发一种客观、有效的方法来衡量重要但主观的技能集,如人际沟通、同理心和高效的数据收集。