Fung Kin Wah, Hole William T, Nelson Stuart J, Srinivasan Suresh, Powell Tammy, Roth Laura
National Library of Medicine, 8600 Rockville Pike, Bethesda, MD 20894, USA.
J Am Med Inform Assoc. 2005 Jul-Aug;12(4):486-94. doi: 10.1197/jamia.M1767. Epub 2005 Mar 31.
The integration of SNOMED CT into the Unified Medical Language System (UMLS) involved the alignment of two views of synonymy that were different because the two vocabulary systems have different intended purposes and editing principles. The UMLS is organized according to one view of synonymy, but its structure also represents all the individual views of synonymy present in its source vocabularies. Despite progress in knowledge-based automation of development and maintenance of vocabularies, manual curation is still the main method of determining synonymy. The aim of this study was to investigate the quality of human judgment of synonymy.
Sixty pairs of potentially controversial SNOMED CT synonyms were reviewed by 11 domain vocabulary experts (six UMLS editors and five noneditors), and scores were assigned according to the degree of synonymy.
The synonymy scores of each subject were compared to the gold standard (the overall mean synonymy score of all subjects) to assess accuracy. Agreement between UMLS editors and noneditors was measured by comparing the mean synonymy scores of editors to noneditors.
Average accuracy was 71% for UMLS editors and 75% for noneditors (difference not statistically significant). Mean scores of editors and noneditors showed significant positive correlation (Spearman's rank correlation coefficient 0.654, two-tailed p < 0.01) with a concurrence rate of 75% and an interrater agreement kappa of 0.43.
The accuracy in the judgment of synonymy was comparable for UMLS editors and nonediting domain experts. There was reasonable agreement between the two groups.
将SNOMED CT整合到统一医学语言系统(UMLS)中涉及两种同义关系视图的对齐,这两种视图不同,因为这两个词汇系统具有不同的预期目的和编辑原则。UMLS是根据一种同义关系视图组织的,但其结构也代表了其源词汇表中存在的所有同义关系的个体视图。尽管在基于知识的词汇表开发和维护自动化方面取得了进展,但人工编纂仍然是确定同义关系的主要方法。本研究的目的是调查同义关系的人工判断质量。
11位领域词汇专家(6位UMLS编辑和5位非编辑)对60对可能存在争议的SNOMED CT同义词进行了审查,并根据同义程度打分。
将每个受试者的同义关系得分与金标准(所有受试者的总体平均同义关系得分)进行比较,以评估准确性。通过比较编辑和非编辑的平均同义关系得分来衡量UMLS编辑和非编辑之间的一致性。
UMLS编辑的平均准确率为71%,非编辑为75%(差异无统计学意义)。编辑和非编辑的平均得分显示出显著的正相关(斯皮尔曼等级相关系数0.654,双侧p<0.01),一致率为75%,评分者间一致性kappa为0.43。
UMLS编辑和非编辑领域专家在同义关系判断上的准确性相当。两组之间有合理的一致性。