Computer Aided Medical Procedures (CAMP), Technische Universität München, Boltzmanstr. 3 Garching, 85748, Germany; Department of Child and Adolescent Psychiatry, Psychosomatic and Psychotherapy, Ludwig-Maximilian-University, Waltherstr. 23. Munich, Germany.
Computer Aided Medical Procedures (CAMP), Technische Universität München, Boltzmanstr. 3 Garching, 85748, Germany.
Med Image Anal. 2017 Oct;41:2-17. doi: 10.1016/j.media.2017.05.002. Epub 2017 May 6.
In this paper, we address the multimodal registration problem from a novel perspective, aiming to predict the transformation aligning images directly from their visual appearance. We formulate the prediction as a supervised regression task, with joint image descriptors as input and the output are the parameters of the transformation that guide the moving image towards alignment. We model the joint local appearance with context aware descriptors that capture both local and global cues simultaneously in the two modalities, while the regression function is based on the gradient boosted trees method capable of handling the very large contextual feature space. The good properties of our predictions allow us to couple them with a simple gradient-based optimization for the final registration. Our approach can be applied to any transformation parametrization as well as a broad range of modality pairs. Our method learns the relationship between the intensity distributions of a pair of modalities by using prior knowledge in the form of a small training set of aligned image pairs (in the order of 1-5 in our experiments). We demonstrate the flexibility and generality of our method by evaluating its performance on a variety of multimodal imaging pairs obtained from two publicly available datasets, RIRE (brain MR, CT and PET) and IXI (brain MR). We also show results for the very challenging deformable registration of Intravascular Ultrasound and Histology images. In these experiments, our approach has a larger capture range when compared to other state-of-the-art methods, while improving registration accuracy in complex cases.
在本文中,我们从一个新的角度解决多模态配准问题,旨在直接从图像的视觉外观预测对齐图像的变换。我们将预测表述为一个监督回归任务,输入是联合图像描述符,输出是引导移动图像对齐的变换参数。我们使用上下文感知描述符来建模联合局部外观,这些描述符同时在两个模态中捕获局部和全局线索,而回归函数基于梯度提升树方法,能够处理非常大的上下文特征空间。我们的预测具有良好的性质,允许我们将它们与基于梯度的简单优化结合起来,以实现最终的配准。我们的方法可以应用于任何变换参数化以及广泛的模态对。我们的方法通过使用小的对齐图像对训练集(在我们的实验中为 1-5 对)的形式将先验知识应用于一对模态的强度分布之间的关系。我们通过评估其在两个公开可用数据集(RIRE(脑 MR、CT 和 PET)和 IXI(脑 MR))获得的各种多模态成像对的性能,展示了我们方法的灵活性和通用性。我们还展示了在血管内超声和组织学图像的非常具有挑战性的可变形配准方面的结果。在这些实验中,与其他最先进的方法相比,我们的方法具有更大的捕获范围,同时在复杂情况下提高了配准精度。