IEEE J Biomed Health Inform. 2019 Nov;23(6):2230-2237. doi: 10.1109/JBHI.2019.2902303. Epub 2019 Feb 28.
The figures found in biomedical literature are a vital part of biomedical research, education, and clinical decision. The multitude of their modalities and the lack of corresponding metadata constitute search and information, retrieval a difficult task. In this paper, we introduce novel multi-label modality classification approaches for biomedical figures without segmenting the compound figures. In particular, we investigate using both simple and compound figures for training a multi-label model to be used for annotating either all figures or only those predicted as compound by a compound figure detection model. Using data from the medical task of ImageCLEF 2016, we train our approaches with visual features and compare them with the approach involving compound figure separation into sub-figures. Furthermore, we study how multimodal learning, from both visual and textual features affects the tasks of classifying biomedical figures by modality and detecting compound figures. Finally, we present a web application for medical figure retrieval, which is based on one of our classification approaches and allows users to search for figures of PubMed Central from any device and provide feedback about the modality of a figure classified by the system.
生物医学文献中的图像是生物医学研究、教育和临床决策的重要组成部分。它们的模态种类繁多,而相应的元数据却缺乏,这使得搜索和信息检索成为一项艰巨的任务。在本文中,我们引入了新颖的生物医学图像多标签模态分类方法,无需对复合图像进行分割。具体来说,我们研究了使用简单图像和复合图像来训练多标签模型,该模型可用于标注所有图像或仅标注复合图像检测模型预测为复合的图像。我们使用来自 2016 年医学图像 CLEF 任务的数据,使用视觉特征训练我们的方法,并将其与涉及将复合图像分离成子图像的方法进行比较。此外,我们研究了多模态学习(来自视觉和文本特征)如何影响通过模态对生物医学图像进行分类和检测复合图像的任务。最后,我们展示了一个基于我们的分类方法之一的医学图像检索的网络应用程序,它允许用户从任何设备搜索 PubMed Central 的图像,并为系统分类的图像的模态提供反馈。