Feng Yidan, Deng Sen, Lyu Jun, Cai Jing, Wei Mingqiang, Qin Jing
IEEE Trans Med Imaging. 2025 Jan;44(1):373-383. doi: 10.1109/TMI.2024.3445969. Epub 2025 Jan 2.
In multi-modal magnetic resonance imaging (MRI), the tasks of imputing or reconstructing the target modality share a common obstacle: the accurate modeling of fine-grained inter-modal differences, which has been sparingly addressed in current literature. These differences stem from two sources: 1) spatial misalignment remaining after coarse registration and 2) structural distinction arising from modality-specific signal manifestations. This paper integrates the previously separate research trajectories of cross-modality synthesis (CMS) and multi-contrast super-resolution (MCSR) to address this pervasive challenge within a unified framework. Connected through generalized down-sampling ratios, this unification not only emphasizes their common goal in reducing structural differences, but also identifies the key task distinguishing MCSR from CMS: modeling the structural distinctions using the limited information from the misaligned target input. Specifically, we propose a composite network architecture with several key components: a label correction module to align the coordinates of multi-modal training pairs, a CMS module serving as the base model, an SR branch to handle target inputs, and a difference projection discriminator for structural distinction-centered adversarial training. When training the SR branch as the generator, the adversarial learning is enhanced with distinction-aware incremental modulation to ensure better-controlled generation. Moreover, the SR branch integrates deformable convolutions to address cross-modal spatial misalignment at the feature level. Experiments conducted on three public datasets demonstrate that our approach effectively balances structural accuracy and realism, exhibiting overall superiority in comprehensive evaluations for both tasks over current state-of-the-art approaches. The code is available at https://github.com/papshare/FGDL.
在多模态磁共振成像(MRI)中,对目标模态进行插补或重建的任务面临一个共同障碍:对细粒度模态间差异进行准确建模,而当前文献对此讨论较少。这些差异源于两个方面:1)粗配准后仍存在的空间错位;2)由模态特定信号表现引起的结构差异。本文将跨模态合成(CMS)和多对比度超分辨率(MCSR)这两条先前独立的研究轨迹整合到一个统一框架中,以应对这一普遍挑战。通过广义下采样率进行连接,这种统一不仅强调了它们在减少结构差异方面的共同目标,还确定了区分MCSR和CMS的关键任务:利用未对齐目标输入中的有限信息对结构差异进行建模。具体而言,我们提出了一种具有几个关键组件的复合网络架构:一个用于对齐多模态训练对坐标的标签校正模块、一个作为基础模型的CMS模块、一个用于处理目标输入的超分辨率(SR)分支以及一个用于以结构差异为中心的对抗训练的差异投影判别器。在将SR分支训练为生成器时,通过区分感知增量调制增强对抗学习,以确保更好地控制生成。此外,SR分支集成了可变形卷积,以在特征层面解决跨模态空间错位问题。在三个公共数据集上进行的实验表明,我们的方法有效地平衡了结构准确性和逼真度,在这两项任务的综合评估中总体上优于当前的最先进方法。代码可在https://github.com/papshare/FGDL获取。