IEEE Trans Med Imaging. 2020 Sep;39(9):2904-2919. doi: 10.1109/TMI.2020.2980117. Epub 2020 Mar 11.
Vascular tree disentanglement and vessel type classification are two crucial steps of the graph-based method for retinal artery-vein (A/V) separation. Existing approaches treat them as two independent tasks and mostly rely on ad hoc rules (e.g. change of vessel directions) and hand-crafted features (e.g. color, thickness) to handle them respectively. However, we argue that the two tasks are highly correlated and should be handled jointly since knowing the A/V type can unravel those highly entangled vascular trees, which in turn helps to infer the types of connected vessels that are hard to classify based on only appearance. Therefore, designing features and models isolatedly for the two tasks often leads to a suboptimal solution of A/V separation. In view of this, this paper proposes a multi-task siamese network which aims to learn the two tasks jointly and thus yields more robust deep features for accurate A/V separation. Specifically, we first introduce Convolution Along Vessel (CAV) to extract the visual features by convolving a fundus image along vessel segments, and the geometric features by tracking the directions of blood flow in vessels. The siamese network is then trained to learn multiple tasks: i) classifying A/V types of vessel segments using visual features only, and ii) estimating the similarity of every two connected segments by comparing their visual and geometric features in order to disentangle the vasculature into individual vessel trees. Finally, the results of two tasks mutually correct each other to accomplish final A/V separation. Experimental results demonstrate that our method can achieve accuracy values of 94.7%, 96.9%, and 94.5% on three major databases (DRIVE, INSPIRE, WIDE) respectively, which outperforms recent state-of-the-arts.
血管树分离和血管类型分类是基于图的视网膜动脉-静脉(A/V)分离方法的两个关键步骤。现有的方法将它们视为两个独立的任务,主要依赖于特定的规则(例如血管方向的变化)和手工制作的特征(例如颜色、厚度)分别处理它们。然而,我们认为这两个任务是高度相关的,应该联合处理,因为知道 A/V 类型可以解开那些高度缠绕的血管树,这反过来又有助于推断仅根据外观难以分类的连接血管的类型。因此,为这两个任务分别设计特征和模型往往会导致 A/V 分离的次优解决方案。鉴于此,本文提出了一种多任务连体网络,旨在联合学习这两个任务,从而为准确的 A/V 分离生成更稳健的深度特征。具体来说,我们首先引入沿血管卷积(CAV),通过沿血管段卷积眼底图像来提取视觉特征,并通过跟踪血管中血流的方向来提取几何特征。然后,使用连体网络学习多个任务:i)仅使用视觉特征对血管段的 A/V 类型进行分类,以及 ii)通过比较它们的视觉和几何特征来估计每两个连接段的相似性,以将脉管系统分离成单个血管树。最后,两个任务的结果相互纠正,以完成最终的 A/V 分离。实验结果表明,我们的方法在三个主要数据库(DRIVE、INSPIRE、WIDE)上分别可以达到 94.7%、96.9%和 94.5%的准确率,优于最新的技术水平。