National Center for Biotechnology Information, 8600 Rockville Pike, Bethesda, Maryland 20894, USA.
BMC Bioinformatics. 2011 Oct 3;12 Suppl 8(Suppl 8):S2. doi: 10.1186/1471-2105-12-S8-S2.
We report the Gene Normalization (GN) challenge in BioCreative III where participating teams were asked to return a ranked list of identifiers of the genes detected in full-text articles. For training, 32 fully and 500 partially annotated articles were prepared. A total of 507 articles were selected as the test set. Due to the high annotation cost, it was not feasible to obtain gold-standard human annotations for all test articles. Instead, we developed an Expectation Maximization (EM) algorithm approach for choosing a small number of test articles for manual annotation that were most capable of differentiating team performance. Moreover, the same algorithm was subsequently used for inferring ground truth based solely on team submissions. We report team performance on both gold standard and inferred ground truth using a newly proposed metric called Threshold Average Precision (TAP-k).
We received a total of 37 runs from 14 different teams for the task. When evaluated using the gold-standard annotations of the 50 articles, the highest TAP-k scores were 0.3297 (k=5), 0.3538 (k=10), and 0.3535 (k=20), respectively. Higher TAP-k scores of 0.4916 (k=5, 10, 20) were observed when evaluated using the inferred ground truth over the full test set. When combining team results using machine learning, the best composite system achieved TAP-k scores of 0.3707 (k=5), 0.4311 (k=10), and 0.4477 (k=20) on the gold standard, representing improvements of 12.4%, 21.8%, and 26.6% over the best team results, respectively.
By using full text and being species non-specific, the GN task in BioCreative III has moved closer to a real literature curation task than similar tasks in the past and presents additional challenges for the text mining community, as revealed in the overall team results. By evaluating teams using the gold standard, we show that the EM algorithm allows team submissions to be differentiated while keeping the manual annotation effort feasible. Using the inferred ground truth we show measures of comparative performance between teams. Finally, by comparing team rankings on gold standard vs. inferred ground truth, we further demonstrate that the inferred ground truth is as effective as the gold standard for detecting good team performance.
我们报告了 BioCreative III 中的基因标准化(GN)挑战,要求参赛团队返回在全文文章中检测到的基因标识符的排名列表。在训练中,准备了 32 篇全文和 500 篇部分注释的文章。共有 507 篇文章被选为测试集。由于注释成本很高,对于所有测试文章,都无法获得黄金标准的人类注释。相反,我们开发了一种期望最大化(EM)算法方法,用于选择少数最能区分团队表现的测试文章进行手动注释。此外,还使用相同的算法仅根据团队提交结果进行了基于事实的推断。我们使用新提出的称为阈值平均精度(TAP-k)的度量标准报告了基于黄金标准和基于推断的事实的团队表现。
我们共收到来自 14 个不同团队的 37 个参赛作品。在使用 50 篇文章的黄金标准注释进行评估时,最高的 TAP-k 分数分别为 0.3297(k=5)、0.3538(k=10)和 0.3535(k=20)。当使用整个测试集的推断事实进行评估时,观察到更高的 TAP-k 分数为 0.4916(k=5、10、20)。使用机器学习组合团队结果时,最佳组合系统在黄金标准上获得的 TAP-k 分数分别为 0.3707(k=5)、0.4311(k=10)和 0.4477(k=20),分别比最佳团队结果提高了 12.4%、21.8%和 26.6%。
通过使用全文且不针对特定物种,BioCreative III 中的 GN 任务比过去的类似任务更接近真实的文献整理任务,并为文本挖掘社区带来了额外的挑战,这在整体团队表现中得到了体现。通过使用黄金标准评估团队,我们表明 EM 算法允许区分团队提交结果,同时保持手动注释的可行性。通过使用推断的事实,我们展示了团队之间的比较性能的度量标准。最后,通过比较黄金标准与推断事实的团队排名,我们进一步证明了推断的事实与黄金标准一样有效,可用于检测优秀的团队表现。