Office of Education Research, Mayo Clinic College of Medicine, Rochester, Minnesota 55905, USA.
Acad Med. 2013 Jun;88(6):872-83. doi: 10.1097/ACM.0b013e31828ffdcf.
To summarize the tool characteristics, sources of validity evidence, methodological quality, and reporting quality for studies of technology-enhanced simulation-based assessments for health professions learners.
The authors conducted a systematic review, searching MEDLINE, EMBASE, CINAHL, ERIC, PsychINFO, Scopus, key journals, and previous reviews through May 2011. They selected original research in any language evaluating simulation-based assessment of practicing and student physicians, nurses, and other health professionals. Reviewers working in duplicate evaluated validity evidence using Messick's five-source framework; methodological quality using the Medical Education Research Study Quality Instrument and the revised Quality Assessment of Diagnostic Accuracy Studies; and reporting quality using the Standards for Reporting Diagnostic Accuracy and Guidelines for Reporting Reliability and Agreement Studies.
Of 417 studies, 350 (84%) involved physicians at some stage in training. Most focused on procedural skills, including minimally invasive surgery (N=142), open surgery (81), and endoscopy (67). Common elements of validity evidence included relations with trainee experience (N=306), content (142), relations with other measures (128), and interrater reliability (124). Of the 217 studies reporting more than one element of evidence, most were judged as having high or unclear risk of bias due to selective sampling (N=192) or test procedures (132). Only 64% proposed a plan for interpreting the evidence to be presented (validity argument).
Validity evidence for simulation-based assessments is sparse and is concentrated within specific specialties, tools, and sources of validity evidence. The methodological and reporting quality of assessment studies leaves much room for improvement.
总结健康职业学习者技术增强模拟评估研究的工具特征、有效性证据来源、方法学质量和报告质量。
作者进行了系统评价,检索了 MEDLINE、EMBASE、CINAHL、ERIC、PsychINFO、Scopus、主要期刊和 2011 年 5 月之前的综述。他们选择了评估执业医师和医学生、护士和其他卫生专业人员模拟评估的任何语言的原始研究。评审员以重复的方式使用 Messick 的五源框架评估有效性证据;使用医学教育研究质量工具和修订的诊断准确性研究质量评估以及报告诊断准确性的标准和报告可靠性和一致性研究的指南评估方法学质量。
在 417 项研究中,350 项(84%)涉及处于培训某个阶段的医生。大多数研究都集中在程序技能上,包括微创外科(N=142)、开放手术(N=81)和内窥镜检查(N=67)。有效性证据的常见要素包括与学员经验(N=306)、内容(N=142)、与其他措施的关系(N=128)和评分者间可靠性(N=124)的关系。在报告了多个证据要素的 217 项研究中,由于选择性采样(N=192)或测试程序(N=132),大多数被认为具有高或不清楚的偏倚风险。只有 64%的研究提出了解释将要呈现的证据的计划(有效性论证)。
基于模拟的评估的有效性证据很少,且集中在特定的专业、工具和有效性证据来源中。评估研究的方法学和报告质量还有很大的改进空间。