Centre for Medical Image Computing, Department of Medical Physics and Biomedical Engineering, University College London, London, UK; Institute of Biomedical Engineering, Department of Engineering Science, University of Oxford, Oxford, UK.
Centre for Medical Image Computing, Department of Medical Physics and Biomedical Engineering, University College London, London, UK; Wellcome / EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK.
Med Image Anal. 2018 Oct;49:1-13. doi: 10.1016/j.media.2018.07.002. Epub 2018 Jul 4.
One of the fundamental challenges in supervised learning for multimodal image registration is the lack of ground-truth for voxel-level spatial correspondence. This work describes a method to infer voxel-level transformation from higher-level correspondence information contained in anatomical labels. We argue that such labels are more reliable and practical to obtain for reference sets of image pairs than voxel-level correspondence. Typical anatomical labels of interest may include solid organs, vessels, ducts, structure boundaries and other subject-specific ad hoc landmarks. The proposed end-to-end convolutional neural network approach aims to predict displacement fields to align multiple labelled corresponding structures for individual image pairs during the training, while only unlabelled image pairs are used as the network input for inference. We highlight the versatility of the proposed strategy, for training, utilising diverse types of anatomical labels, which need not to be identifiable over all training image pairs. At inference, the resulting 3D deformable image registration algorithm runs in real-time and is fully-automated without requiring any anatomical labels or initialisation. Several network architecture variants are compared for registering T2-weighted magnetic resonance images and 3D transrectal ultrasound images from prostate cancer patients. A median target registration error of 3.6 mm on landmark centroids and a median Dice of 0.87 on prostate glands are achieved from cross-validation experiments, in which 108 pairs of multimodal images from 76 patients were tested with high-quality anatomical labels.
在监督学习的多模态图像配准中,一个基本的挑战是缺乏体素级空间对应关系的真实数据。本工作描述了一种从包含在解剖学标签中的高级对应信息推断体素级变换的方法。我们认为,与体素级对应相比,这种标签对于图像对的参考集来说更可靠和实用。典型的感兴趣的解剖标签可能包括实体器官、血管、导管、结构边界和其他特定于主体的临时地标。所提出的端到端卷积神经网络方法旨在通过在训练期间对齐多个标记的对应结构来预测位移场,而仅使用未标记的图像对作为网络输入进行推断。我们强调了所提出策略的通用性,用于训练,利用各种类型的解剖标签,这些标签不需要在所有训练图像对上都可识别。在推断时,所得到的 3D 可变形图像配准算法实时运行,完全自动化,不需要任何解剖标签或初始化。比较了几种网络架构变体,用于注册来自前列腺癌患者的 T2 加权磁共振图像和 3D 经直肠超声图像。在交叉验证实验中,从 76 名患者的 108 对多模态图像中测试了具有高质量解剖标签的情况下,在地标质心处实现了 3.6mm 的平均目标注册误差,在前列腺中实现了 0.87 的平均 Dice 系数。