Elkin Peter L, Brown Steven H, Bauer Brent A, Husser Casey S, Carruth William, Bergstrom Larry R, Wahner-Roedler Dietlind L
Department of Medicine, Mayo Foundation, Rochester, MN, USA.
BMC Med Inform Decis Mak. 2005 May 5;5:13. doi: 10.1186/1472-6947-5-13.
Identification of negation in electronic health records is essential if we are to understand the computable meaning of the records: Our objective is to compare the accuracy of an automated mechanism for assignment of Negation to clinical concepts within a compositional expression with Human Assigned Negation. Also to perform a failure analysis to identify the causes of poorly identified negation (i.e. Missed Conceptual Representation, Inaccurate Conceptual Representation, Missed Negation, Inaccurate identification of Negation).
41 Clinical Documents (Medical Evaluations; sometimes outside of Mayo these are referred to as History and Physical Examinations) were parsed using the Mayo Vocabulary Server Parsing Engine. SNOMED-C was used to provide concept coverage for the clinical concepts in the record. These records resulted in identification of Concepts and textual clues to Negation. These records were reviewed by an independent medical terminologist, and the results were tallied in a spreadsheet. Where questions on the review arose Internal Medicine Faculty were employed to make a final determination.
SNOMED-CT was used to provide concept coverage of the 14,792 Concepts in 41 Health Records from John's Hopkins University. Of these, 1,823 Concepts were identified as negative by Human review. The sensitivity (Recall) of the assignment of negation was 97.2% (p < 0.001, Pearson Chi-Square test; when compared to a coin flip). The specificity of assignment of negation was 98.8%. The positive likelihood ratio of the negation was 81. The positive predictive value (Precision) was 91.2%
Automated assignment of negation to concepts identified in health records based on review of the text is feasible and practical. Lexical assignment of negation is a good test of true Negativity as judged by the high sensitivity, specificity and positive likelihood ratio of the test. SNOMED-CT had overall coverage of 88.7% of the concepts being negated.
如果我们想要理解电子健康记录的可计算含义,那么识别其中的否定信息至关重要。我们的目标是比较一种将否定信息自动分配给组合表达式中临床概念的机制与人工分配否定信息的准确性。同时进行失败分析,以确定否定信息识别不佳的原因(即概念表示遗漏、概念表示不准确、否定遗漏、否定识别不准确)。
使用梅奥词汇服务器解析引擎对41份临床文档(医学评估;在梅奥之外,这些有时被称为病史和体格检查)进行解析。使用SNOMED-C为记录中的临床概念提供概念覆盖。这些记录导致了概念的识别和否定的文本线索。这些记录由一名独立的医学术语专家进行审查,结果记录在电子表格中。当审查中出现问题时,会聘请内科教员做出最终决定。
使用SNOMED-CT为约翰霍普金斯大学41份健康记录中的14792个概念提供概念覆盖。其中,1823个概念经人工审查被确定为否定。否定分配的敏感性(召回率)为97.2%(p<0.001,Pearson卡方检验;与抛硬币相比)。否定分配的特异性为98.8%。否定的阳性似然比为81。阳性预测值(精确率)为91.2%。
基于文本审查将否定信息自动分配给健康记录中识别出的概念是可行且实用的。根据该测试的高敏感性、特异性和阳性似然比判断,否定的词汇分配是对真正否定性的良好测试。SNOMED-CT对被否定概念的总体覆盖率为88.7%。