Scully Rebecca E, Deal Shanley B, Clark Michael J, Yang Katherine, Wnuk Greg, Smink Douglas S, Fryer Jonathan P, Bohnen Jordan D, Teitelbaum Ezra N, Meyerson Shari L, Meier Andreas H, Gauger Paul G, Reddy Rishindra M, Kendrick Daniel E, Stern Michael, Hughes David T, Chipman Jeffrey G, Patel Jitesh A, Alseidi Adnan, George Brian C
Department of Surgery, Brigham and Women's Hospital, Boston, Massachusetts.
Department of Surgery, Virginia Mason Medical Center, Seattle, Washington.
J Surg Educ. 2020 May-Jun;77(3):627-634. doi: 10.1016/j.jsurg.2019.12.016. Epub 2020 Mar 20.
We examined the impact of video editing and rater expertise in surgical resident evaluation on operative performance ratings of surgical trainees.
Randomized independent review of intraoperative video.
Operative video was captured at a single, tertiary hospital in Boston, MA.
Six common general surgery procedures were video recorded of 6 attending-trainee dyads. Full-length and condensed versions (n = 12 videos) were then reviewed by 13 independent surgeon raters (5 evaluation experts, 8 nonexperts) using a crossed design. Trainee performance was rated using the Operative Performance Rating Scale, System for Improving and Measuring Procedural Learning (SIMPL) Performance scale, the Zwisch scale, and ten Cate scale. These ratings were then standardized before being compared using Bayesian mixed models with raters and videos treated as random effects.
Editing had no effect on the Operative Performance Rating Scale Overall Performance (-0.10, p = 0.30), SIMPL Performance (0.13, p = 0.71), Zwisch (-0.12, p = 0.27), and ten Cate scale (-0.13, p = 0.29). Additionally, rater expertise (evaluation expert vs. nonexpert) had no effect on the same scales (-0.16 (p = 0.32), 0.18 (p = 0.74), 0.25 (p = 0.81), and 0.25 (p = 0.17).
There is little difference in operative performance assessment scores when raters use condensed videos or when raters who are not experts in surgical resident evaluation are used. Future validation studies of operative performance assessment scales may be facilitated by using nonexpert surgeon raters viewing videos condensed using a standardized protocol.
我们研究了视频编辑和评估者专业知识对外科住院医师评估中手术学员手术操作评分的影响。
术中视频的随机独立评审。
手术视频在马萨诸塞州波士顿的一家三级医院采集。
对6对主刀医生-学员组合进行了6种常见普通外科手术的视频录制。然后,13名独立的外科医生评估者(5名评估专家、8名非专家)采用交叉设计对全长和精简版(n = 12个视频)进行评审。使用手术操作评分量表、改善和测量程序学习系统(SIMPL)表现量表、兹维施量表和滕凯特量表对学员表现进行评分。然后在将评估者和视频视为随机效应,使用贝叶斯混合模型进行比较之前,对这些评分进行标准化。
编辑对手术操作评分量表总体表现(-0.10,p = 0.30)、SIMPL表现(0.13,p = 0.71)、兹维施量表(-0.12,p = 0.27)和滕凯特量表(-0.13,p = 0.29)没有影响。此外,评估者专业知识(评估专家与非专家)对相同量表也没有影响(-0.16(p = 0.32)、0.18(p = 0.74)、0.25(p = 0.81)和0.25(p = 0.17)。
当评估者使用精简视频或使用非外科住院医师评估专家的评估者时,手术操作评估分数几乎没有差异。使用非专家外科医生评估者观看按照标准化方案精简的视频,可能有助于未来手术操作评估量表的验证研究。