Zhu Wenhui, Qiu Peijie, Chen Xiwen, Li Xin, Lepore Natasha, Dumitrascu Oana M, Wang Yalin
School of Computing and Augmented Intelligence, Arizona State University.
McKeley School of Engineering, Washington University in St. Louis.
Conf Comput Vis Pattern Recognit Workshops. 2024 Jun;2024:2285-2294. doi: 10.1109/CVPRW63382.2024.00234. Epub 2024 Sep 27.
Over the past few decades, convolutional neural networks (CNNs) have been at the forefront of the detection and tracking of various retinal diseases (RD). Despite their success, the emergence of vision transformers (ViT) in the 2020s has shifted the trajectory of RD model development. The leading-edge performance of ViT-based models in RD can be largely credited to their scalability-their ability to improve as more parameters are added. As a result, ViT-based models tend to outshine traditional CNNs in RD applications, albeit at the cost of increased data and computational demands. ViTs also differ from CNNs in their approach to processing images, working with patches rather than local regions, which can complicate the precise localization of small, variably presented lesions in RD. In our study, we revisited and updated the architecture of a CNN model, specifically MobileNet, to enhance its utility in RD diagnostics. We found that an optimized MobileNet, through selective modifications, can surpass ViT-based models in various RD benchmarks, including diabetic retinopathy grading, detection of multiple fundus diseases, and classification of diabetic macular edema. The code is available at https://github.com/Retinal-Research/NN-MOBILENET.
在过去几十年里,卷积神经网络(CNN)一直处于各种视网膜疾病(RD)检测与跟踪的前沿。尽管取得了成功,但2020年代视觉Transformer(ViT)的出现改变了视网膜疾病模型的发展轨迹。基于ViT的模型在视网膜疾病方面的前沿性能很大程度上归功于其可扩展性——随着添加更多参数而提升性能的能力。因此,基于ViT的模型在视网膜疾病应用中往往比传统的CNN更出色,尽管代价是数据和计算需求增加。ViT在处理图像的方式上也与CNN不同,它处理的是图像块而非局部区域,这可能会使视网膜疾病中小的、呈现方式各异的病变的精确定位变得复杂。在我们的研究中,我们重新审视并更新了一个CNN模型(具体来说是MobileNet)的架构,以增强其在视网膜疾病诊断中的效用。我们发现,通过选择性修改而优化的MobileNet,在包括糖尿病视网膜病变分级、多种眼底疾病检测以及糖尿病黄斑水肿分类等各种视网膜疾病基准测试中,可以超越基于ViT的模型。代码可在https://github.com/Retinal-Research/NN-MOBILENET获取。