Wang Weisen, Li Xirong, Xu Zhiyan, Yu Weihong, Zhao Jianchun, Ding Dayong, Chen Youxin
IEEE J Biomed Health Inform. 2022 Aug;26(8):4111-4122. doi: 10.1109/JBHI.2022.3171523. Epub 2022 Aug 11.
This paper tackles automated categorization of Age-related Macular Degeneration (AMD), a common macular disease among people over 50. Previous research efforts mainly focus on AMD categorization with a single-modal input, let it be a color fundus photograph (CFP) or an OCT B-scan image. By contrast, we consider AMD categorization given a multi-modal input, a direction that is clinically meaningful yet mostly unexplored. Contrary to the prior art that takes a traditional approach of feature extraction plus classifier training that cannot be jointly optimized, we opt for end-to-end multi-modal Convolutional Neural Networks (MM-CNN). Our MM-CNN is instantiated by a two-stream CNN, with spatially-invariant fusion to combine information from the CFP and OCT streams. In order to visually interpret the contribution of the individual modalities to the final prediction, we extend the class activation mapping (CAM) technique to the multi-modal scenario. For effective training of MM-CNN, we develop two data augmentation methods. One is GAN-based CFP/OCT image synthesis, with our novel use of CAMs as conditional input of a high-resolution image-to-image translation GAN. The other method is Loose Pairing, which pairs a CFP image and an OCT image on the basis of their classes instead of eye identities. Experiments on a clinical dataset consisting of 1,094 CFP images and 1,289 OCT images acquired from 1,093 distinct eyes show that the proposed solution obtains better F1 and Accuracy than multiple baselines for multi-modal AMD categorization. Code and data are available at https://github.com/li-xirong/mmc-amd.
本文致力于年龄相关性黄斑变性(AMD)的自动分类,AMD是50岁以上人群中常见的黄斑疾病。以往的研究主要集中在基于单模态输入的AMD分类,无论是彩色眼底照片(CFP)还是OCT B扫描图像。相比之下,我们考虑基于多模态输入的AMD分类,这是一个具有临床意义但大多未被探索的方向。与采用传统特征提取加分类器训练方法且无法联合优化的现有技术不同,我们选择端到端的多模态卷积神经网络(MM-CNN)。我们的MM-CNN由双流CNN实例化,通过空间不变融合来组合来自CFP和OCT流的信息。为了直观地解释各个模态对最终预测的贡献,我们将类激活映射(CAM)技术扩展到多模态场景。为了有效训练MM-CNN,我们开发了两种数据增强方法。一种是基于GAN的CFP/OCT图像合成,我们创新性地将CAM用作高分辨率图像到图像翻译GAN的条件输入。另一种方法是宽松配对,它基于CFP图像和OCT图像的类别而不是眼睛标识来进行配对。对一个由从1093只不同眼睛获取的1094张CFP图像和1289张OCT图像组成的临床数据集进行的实验表明,所提出的解决方案在多模态AMD分类方面比多个基线获得了更好的F1值和准确率。代码和数据可在https://github.com/li-xirong/mmc-amd获取。