Inserm, UMR 1101, SFR ScInBioS, F-29200 Brest, France.
Inserm, UMR 1101, SFR ScInBioS, F-29200 Brest, France; Univ Bretagne Occidentale, F-29200 Brest, France.
Med Image Anal. 2016 Apr;29:47-64. doi: 10.1016/j.media.2015.12.006. Epub 2015 Dec 29.
With the increased prevalence of retinal pathologies, automating the detection of these pathologies is becoming more and more relevant. In the past few years, many algorithms have been developed for the automated detection of a specific pathology, typically diabetic retinopathy, using eye fundus photography. No matter how good these algorithms are, we believe many clinicians would not use automatic detection tools focusing on a single pathology and ignoring any other pathology present in the patient's retinas. To solve this issue, an algorithm for characterizing the appearance of abnormal retinas, as well as the appearance of the normal ones, is presented. This algorithm does not focus on individual images: it considers examination records consisting of multiple photographs of each retina, together with contextual information about the patient. Specifically, it relies on data mining in order to learn diagnosis rules from characterizations of fundus examination records. The main novelty is that the content of examination records (images and context) is characterized at multiple levels of spatial and lexical granularity: 1) spatial flexibility is ensured by an adaptive decomposition of composite retinal images into a cascade of regions, 2) lexical granularity is ensured by an adaptive decomposition of the feature space into a cascade of visual words. This multigranular representation allows for great flexibility in automatically characterizing normality and abnormality: it is possible to generate diagnosis rules whose precision and generalization ability can be traded off depending on data availability. A variation on usual data mining algorithms, originally designed to mine static data, is proposed so that contextual and visual data at adaptive granularity levels can be mined. This framework was evaluated in e-ophtha, a dataset of 25,702 examination records from the OPHDIAT screening network, as well as in the publicly-available Messidor dataset. It was successfully applied to the detection of patients that should be referred to an ophthalmologist and also to the specific detection of several pathologies.
随着视网膜病变的发病率不断增加,自动检测这些病变变得越来越重要。在过去的几年中,已经开发出许多算法来使用眼底摄影自动检测特定的病变,通常是糖尿病视网膜病变。无论这些算法有多好,我们相信许多临床医生都不会使用专注于单一病变的自动检测工具,而忽略患者视网膜中存在的任何其他病变。为了解决这个问题,提出了一种用于描述异常视网膜以及正常视网膜外观的算法。该算法不专注于单个图像:它考虑了由每个视网膜的多张照片以及有关患者的上下文信息组成的检查记录。具体来说,它依赖于数据挖掘,以便从眼底检查记录的特征描述中学习诊断规则。主要的新颖之处在于,检查记录的内容(图像和上下文)在多个空间和词汇粒度级别上进行了特征化:1)通过将复合视网膜图像自适应地分解为区域级联,确保了空间灵活性,2)通过将特征空间自适应地分解为视觉单词级联,确保了词汇粒度。这种多粒度表示形式允许在自动描述正常和异常情况时具有很大的灵活性:可以生成诊断规则,其精度和泛化能力可以根据数据可用性进行权衡。提出了一种变体通常用于挖掘静态数据的数据挖掘算法,以便可以挖掘自适应粒度级别的上下文和视觉数据。该框架在 e-ophtha 中进行了评估,这是一个来自 OPHDIAT 筛查网络的 25,702 个检查记录的数据集,以及公开可用的 Messidor 数据集。它成功地应用于检测应转介给眼科医生的患者,以及用于特定的几种病变的检测。