Benvenuto Giovana A, Colnago Marilaine, Dias Maurício A, Negri Rogério G, Silva Erivaldo A, Casaca Wallace
Faculty of Science and Technology (FCT), São Paulo State University (UNESP), Presidente Prudente 19060-900, Brazil.
Institute of Mathematics and Computer Science (ICMC), São Paulo University (USP), São Carlos 13566-590, Brazil.
Bioengineering (Basel). 2022 Aug 5;9(8):369. doi: 10.3390/bioengineering9080369.
In ophthalmology, the registration problem consists of finding a geometric transformation that aligns a pair of images, supporting eye-care specialists who need to record and compare images of the same patient. Considering the registration methods for handling eye fundus images, the literature offers only a limited number of proposals based on deep learning (DL), whose implementations use the supervised learning paradigm to train a model. Additionally, ensuring high-quality registrations while still being flexible enough to tackle a broad range of fundus images is another drawback faced by most existing methods in the literature. Therefore, in this paper, we address the above-mentioned issues by introducing a new DL-based framework for eye fundus registration. Our methodology combines a U-shaped fully convolutional neural network with a spatial transformation learning scheme, where a reference-free similarity metric allows the registration without assuming any pre-annotated or artificially created data. Once trained, the model is able to accurately align pairs of images captured under several conditions, which include the presence of anatomical differences and low-quality photographs. Compared to other registration methods, our approach achieves better registration outcomes by just passing as input the desired pair of fundus images.
在眼科领域,配准问题在于找到一种几何变换,以对齐一对图像,为需要记录和比较同一患者图像的眼科护理专家提供支持。考虑到处理眼底图像的配准方法,文献中基于深度学习(DL)的提议数量有限,其实现使用监督学习范式来训练模型。此外,在保证高质量配准的同时,还要足够灵活以处理各种眼底图像,这是文献中大多数现有方法面临的另一个缺点。因此,在本文中,我们通过引入一种新的基于深度学习的眼底配准框架来解决上述问题。我们的方法将一个U形全卷积神经网络与一种空间变换学习方案相结合,其中一种无参考相似性度量允许在不假设任何预先标注或人工创建数据的情况下进行配准。一旦训练完成,该模型能够准确对齐在多种条件下拍摄的图像对,这些条件包括存在解剖差异和低质量照片。与其他配准方法相比,我们的方法只需将所需的一对眼底图像作为输入,就能获得更好的配准结果。