King Andrew J, Kahn Jeremy M, Brant Emily B, Cooper Gregory F, Mowery Danielle L
Department of Critical Care Medicine.
Department of Biomedical Informatics, University of Pittsburgh, Pittsburgh, Pennsylvania, and.
ATS Sch. 2022 Sep 23;3(4):548-560. doi: 10.34197/ats-scholar.2022-0010OC. eCollection 2022 Dec.
Oral case presentation is a crucial skill of physicians and a key component of team-based care. However, consistent and objective assessment and feedback on presentations during training are infrequent.
To determine the potential value of applying natural language processing, computer software that extracts meaning from text, to transcripts of oral case presentations as a strategy to assess their quality automatically and objectively.
We transcribed a collection of simulated oral case presentations. The presentations were from eight critical care fellows and one critical care attending. They were instructed to review the medical charts of 11 real intensive care unit patient cases and to audio record themselves, presenting each case as if they were doing so on morning rounds. We then used natural language processing to convert the transcripts from human-readable text into machine-readable numbers. These numbers represent details of the presentation style and content. The distance between the numeric representation of two different transcripts negatively correlates with the similarity of those two transcripts. We ranked fellows on the basis of how similar their presentations were to the attending's presentations.
The 99 presentations included 260 minutes of audio (mean length: 2.6 ± 1.24 min per case). On average, 23.88 ± 2.65 sentences were spoken, and each sentence had 14.10 ± 0.67 words, 3.62 ± 0.15 medical concepts, and 0.75 ± 0.09 medical adjectives. When ranking fellows on the basis of how similar their presentations were to the attending's presentation, we found a gap between the five fellows with the most similar presentations and the three fellows with the least similar presentations (average group similarity scores of 0.62 ± 0.01 and 0.53 ± 0.01, respectively). Rankings were sensitive to whether presentation style or content information were weighted more heavily when calculating transcript similarity.
Natural language processing enabled the ranking of case presentations on the basis of how similar they were to a reference presentation. Although additional work is needed to convert these rankings, and underlying similarity scores, into actionable feedback for trainees, these methods may support new tools for improving medical education.
口头病例汇报是医生的一项关键技能,也是团队医疗的重要组成部分。然而,在培训期间,对病例汇报进行持续且客观的评估和反馈并不常见。
确定应用自然语言处理(从文本中提取含义的计算机软件)对口头病例汇报的文字记录进行分析,以此作为自动、客观评估汇报质量的一种策略的潜在价值。
我们转录了一组模拟口头病例汇报。这些汇报来自八位重症医学研究员和一位重症医学主治医师。他们被要求查阅11例真实重症监护病房患者的病历,并进行录音,就像在早查房时那样汇报每个病例。然后,我们使用自然语言处理将文字记录从人类可读文本转换为机器可读数字。这些数字代表了汇报风格和内容的细节。两份不同文字记录的数字表示之间的距离与这两份文字记录的相似度呈负相关。我们根据研究员的汇报与主治医师汇报的相似程度对他们进行排名。
99份汇报包含260分钟音频(平均时长:每个病例2.6 ± 1.24分钟)。平均而言,每份汇报说了23.88 ± 2.65个句子,每个句子有14.10 ± 0.67个单词、3.62 ± 0.15个医学概念和0.75 ± 0.09个医学形容词。当根据研究员的汇报与主治医师汇报的相似程度对他们进行排名时,我们发现五份最相似汇报的研究员与三份最不相似汇报的研究员之间存在差距(平均组相似度得分分别为0.62 ± 0.01和0.53 ± 0.01)。在计算文字记录相似度时,排名对汇报风格或内容信息的权重更为敏感。
自然语言处理能够根据病例汇报与参考汇报的相似程度对其进行排名。尽管需要开展更多工作将这些排名以及潜在的相似度得分转化为针对实习生的可操作反馈,但这些方法可能会为改进医学教育提供新工具。