Hua Cam-Hao, Huynh-The Thien, Lee Sungyoung
Annu Int Conf IEEE Eng Med Biol Soc. 2020 Jul;2020:1992-1995. doi: 10.1109/EMBC44109.2020.9175355.
Diabetic Retinopathy (DR), the complication leading to vision loss, is generally graded according to the amalgamation of various structural factors in fundus photography such as number of microaneurysms, hemorrhages, vascular abnormalities, etc. To this end, Convolution Neural Network (CNN) with impressively representational power has been exhaustively utilized to address this problem. However, while existing multi-stream networks are costly, the conventional CNNs do not consider multiple levels of semantic context, which suffers from the loss of spatial correlations between the aforementioned DR-related signs. Therefore, this paper proposes a Densely Reversed Attention based CNN (DRAN) to leverage the learnable integration of channel-wise attention at multi-level features in a pretrained network for unambiguously involving spatial representations of important DR-oriented factors. Consequently, the proposed approach gains a quadratic weighted kappa of 85.6% on Kaggle DR detection dataset, which is competitive with the state-of-the-arts.
糖尿病视网膜病变(DR)是导致视力丧失的并发症,通常根据眼底照片中各种结构因素的综合情况进行分级,如微动脉瘤数量、出血、血管异常等。为此,具有强大表征能力的卷积神经网络(CNN)已被广泛用于解决这一问题。然而,现有的多流网络成本高昂,而传统的CNN没有考虑多层次的语义上下文,存在上述与DR相关体征之间空间相关性丢失的问题。因此,本文提出了一种基于密集反向注意力的CNN(DRAN),以利用预训练网络中多层次特征上通道注意力的可学习整合,明确纳入重要的面向DR因素的空间表示。结果,该方法在Kaggle DR检测数据集上获得了85.6%的二次加权卡帕值,与当前最先进的方法具有竞争力。