Wei Chih-Hsuan, Peng Yifan, Leaman Robert, Davis Allan Peter, Mattingly Carolyn J, Li Jiao, Wiegers Thomas C, Lu Zhiyong
National Center for Biotechnology Information, Bethesda, MD 20894, USA.
National Center for Biotechnology Information, Bethesda, MD 20894, USA Department of Computer and Information Sciences, University of Delaware, Newark, DE 19716, USA.
Database (Oxford). 2016 Mar 19;2016. doi: 10.1093/database/baw032. Print 2016.
Manually curating chemicals, diseases and their relationships is significantly important to biomedical research, but it is plagued by its high cost and the rapid growth of the biomedical literature. In recent years, there has been a growing interest in developing computational approaches for automatic chemical-disease relation (CDR) extraction. Despite these attempts, the lack of a comprehensive benchmarking dataset has limited the comparison of different techniques in order to assess and advance the current state-of-the-art. To this end, we organized a challenge task through BioCreative V to automatically extract CDRs from the literature. We designed two challenge tasks: disease named entity recognition (DNER) and chemical-induced disease (CID) relation extraction. To assist system development and assessment, we created a large annotated text corpus that consisted of human annotations of chemicals, diseases and their interactions from 1500 PubMed articles. 34 teams worldwide participated in the CDR task: 16 (DNER) and 18 (CID). The best systems achieved an F-score of 86.46% for the DNER task--a result that approaches the human inter-annotator agreement (0.8875)--and an F-score of 57.03% for the CID task, the highest results ever reported for such tasks. When combining team results via machine learning, the ensemble system was able to further improve over the best team results by achieving 88.89% and 62.80% in F-score for the DNER and CID task, respectively. Additionally, another novel aspect of our evaluation is to test each participating system's ability to return real-time results: the average response time for each team's DNER and CID web service systems were 5.6 and 9.3 s, respectively. Most teams used hybrid systems for their submissions based on machining learning. Given the level of participation and results, we found our task to be successful in engaging the text-mining research community, producing a large annotated corpus and improving the results of automatic disease recognition and CDR extraction. Database URL: http://www.biocreative.org/tasks/biocreative-v/track-3-cdr/.
人工整理化学物质、疾病及其关系对生物医学研究极为重要,但却面临成本高昂以及生物医学文献快速增长的困扰。近年来,人们对开发用于自动提取化学-疾病关系(CDR)的计算方法兴趣日增。尽管有这些尝试,但缺乏全面的基准数据集限制了不同技术之间的比较,从而无法评估和推进当前的技术水平。为此,我们通过生物创意V组织了一项挑战任务,以从文献中自动提取CDR。我们设计了两项挑战任务:疾病命名实体识别(DNER)和化学诱导疾病(CID)关系提取。为协助系统开发和评估,我们创建了一个大型注释文本语料库,该语料库包含来自1500篇PubMed文章的化学物质、疾病及其相互作用的人工注释。全球34个团队参与了CDR任务:16个团队参与DNER任务,18个团队参与CID任务。最佳系统在DNER任务中取得了86.46%的F值——这一结果接近人工注释者之间的一致性(0.8875)——在CID任务中取得了57.03%的F值,这是此类任务中报告的最高结果。当通过机器学习合并各团队结果时,集成系统在DNER和CID任务的F值方面分别达到88.89%和62.80%,能够在最佳团队结果的基础上进一步提高。此外,我们评估的另一个新颖之处在于测试每个参与系统返回实时结果的能力:各团队DNER和CID网络服务系统的平均响应时间分别为5.6秒和9.3秒。大多数团队在提交作品时使用了基于机器学习的混合系统。鉴于参与程度和结果,我们发现我们的任务成功地吸引了文本挖掘研究社区,产生了一个大型注释语料库,并提高了自动疾病识别和CDR提取的结果。数据库网址:http://www.biocreative.org/tasks/biocreative-v/track-3-cdr/