Department of Veterinary Clinical Medicine, University of Illinois, Urbana, IL.
Department of Mechanical Science and Engineering, University of Illinois, Urbana, IL.
Am J Vet Res. 2023 Jun 27;84(8). doi: 10.2460/ajvr.23.03.0058. Print 2023 Aug 1.
To evaluate suturing skills of veterinary students using 3 common performance assessments (PAs) and to compare findings to data obtained by an electromyographic armband.
16 second-year veterinary students.
Students performed 4 suturing tasks on synthetic tissue models 1, 3, and 5 weeks after a surgical skills course. Digital videos were scored by 4 expert surgeons using 3 PAs (an Objective Structured Clinical Examination [OSCE]- style surgical binary checklist, an Objective Structured Assessment of Technical Skill [OSATS] checklist, and a surgical Global Rating Scale [GRS]). Surface electromyography (sEMG) data collected from the dominant forearm were input to machine learning algorithms. Performance assessment scores were compared between experts and correlated to task completion times and sEMG data. Inter-rater reliability was calculated using the intraclass correlation coefficient (ICC). Inter-rater agreement was calculated using percent agreement with varying levels of tolerance.
Reliability was moderate for the OSCE and OSATS checklists and poor for the GRS. Agreement was achieved for the checklists when moderate tolerance was applied but remained poor for the GRS. sEMG signals did not correlate well with checklist scores or task times, but features extracted from signals permitted task differentiation by routine statistical comparison and correct task classification using machine learning algorithms.
Reliability and agreement of an OSCE-style checklist, OSATS checklist, and surgical GRS assessment were insufficient to characterize suturing skills of veterinary students. To avoid subjectivity associated with PA by raters, further study of kinematics and EMG data is warranted in the surgical skills evaluation of veterinary students.
使用 3 种常见的绩效评估(PA)评估兽医学生的缝合技能,并将评估结果与肌电臂带获得的数据进行比较。
16 名二年级兽医学生。
学生在外科技能课程后 1、3 和 5 周在合成组织模型上进行 4 项缝合任务。数字视频由 4 名专家外科医生使用 3 种 PA(客观结构化临床考试[OSCE]风格的手术二进制检查表、客观结构化评估技术技能[OSATS]检查表和手术综合评分[GRS])进行评分。从主导前臂采集的表面肌电图(sEMG)数据被输入机器学习算法。在专家之间比较绩效评估得分,并将其与任务完成时间和 sEMG 数据相关联。使用组内相关系数(ICC)计算评分者间可靠性。使用不同容忍度的百分比一致性来计算评分者间的一致性。
OSCE 和 OSATS 检查表的可靠性为中度,GRS 的可靠性为较差。当应用中等容忍度时,检查表的一致性达成,但 GRS 的一致性仍然较差。sEMG 信号与检查表得分或任务时间相关性不佳,但从信号中提取的特征可通过常规统计比较区分任务,并使用机器学习算法正确分类任务。
OSCE 风格检查表、OSATS 检查表和手术 GRS 评估的可靠性和一致性不足以描述兽医学生的缝合技能。为了避免评分者的 PA 主观性,需要进一步研究兽医学生手术技能评估中的运动学和肌电图数据。