Ginsburg Shiphra, Regehr Glenn, Mylopoulos Maria
Department of Medicine, Faculty of Medicine, University of Toronto, Toronto, Ontario, Canada.
Med Educ. 2009 May;43(5):414-25. doi: 10.1111/j.1365-2923.2009.03335.x.
This study aimed to explore faculty attendings' scoring and opinions of students' written responses to professionally challenging situations.
In this mixed-methods study, 10 pairs of faculty attendings (attending physicians in internal medicine) marked responses to a professionalism written examination taken by 40 medical students and were then interviewed regarding their scoring decisions. Quantitatively, inter-rater scoring agreement was calculated for each pair and students' global scores were compared with a previously developed theoretical framework. Qualitatively, interviews were analysed using grounded theory.
Inter-rater reliability in scoring was poor. There was also no correlation between faculty's scores and our previous theoretical framework; this lack of correlation persisted despite modifications to the framework. Qualitative analysis of faculty attendings' interviews yielded three major themes: faculty preferred responses in which students expressed insight, showed responsibility, and ultimately put the patient first. Faculty also expressed difficulty in deciding what was more important (the behaviour or the rationale behind it) and in assigning numerical scores to students' responses. Interestingly, they did not downgrade students for mentioning implications for themselves as long as these were balanced by other considerations.
This study attempted to overcome some of the instability that results when we judge behaviours by making the rationales behind students' behaviours explicit. However, between-faculty agreement was still poor. This reinforces concerns that professionalism, as a subtle and complex construct, does not reduce easily to numerical scales. Instead of concentrating on creating the 'perfect' evaluation instrument, educators should perhaps begin to explore alternative approaches, including those that do not rely on numerical scales.
本研究旨在探讨指导教师对学生针对专业挑战性情境的书面回答的评分及看法。
在这项混合方法研究中,10对指导教师(内科主治医师)对40名医学生参加的一次职业素养书面考试的回答进行评分,随后就他们的评分决定接受访谈。在定量方面,计算每对教师之间的评分一致性,并将学生的总体分数与先前制定的理论框架进行比较。在定性方面,使用扎根理论对访谈进行分析。
评分者间的评分信度较差。教师的评分与我们先前的理论框架之间也没有相关性;尽管对该框架进行了修改,这种缺乏相关性的情况仍然存在。对指导教师访谈的定性分析产生了三个主要主题:教师更喜欢学生表达见解、展现责任感并最终将患者放在首位的回答。教师在决定什么更重要(行为还是其背后的理由)以及为学生的回答给出数字评分时也表示困难。有趣的是,只要学生提及对自身的影响能被其他考量因素平衡,教师就不会因此给学生扣分。
本研究试图通过明确学生行为背后的理由来克服我们在评判行为时产生的一些不稳定性。然而,教师之间的一致性仍然较差。这强化了这样一种担忧,即职业素养作为一种微妙而复杂的概念,不容易简化为数字量表。教育工作者或许不应专注于创建“完美”的评估工具,而应开始探索其他方法,包括那些不依赖数字量表的方法。