IEEE Trans Med Imaging. 2024 Feb;43(2):745-759. doi: 10.1109/TMI.2023.3320812. Epub 2024 Feb 2.
Low-dose computed tomography (CT) images suffer from noise and artifacts due to photon starvation and electronic noise. Recently, some works have attempted to use diffusion models to address the over-smoothness and training instability encountered by previous deep-learning-based denoising models. However, diffusion models suffer from long inference time due to a large number of sampling steps involved. Very recently, cold diffusion model generalizes classical diffusion models and has greater flexibility. Inspired by cold diffusion, this paper presents a novel COntextual eRror-modulated gEneralized Diffusion model for low-dose CT (LDCT) denoising, termed CoreDiff. First, CoreDiff utilizes LDCT images to displace the random Gaussian noise and employs a novel mean-preserving degradation operator to mimic the physical process of CT degradation, significantly reducing sampling steps thanks to the informative LDCT images as the starting point of the sampling process. Second, to alleviate the error accumulation problem caused by the imperfect restoration operator in the sampling process, we propose a novel ContextuaL Error-modulAted Restoration Network (CLEAR-Net), which can leverage contextual information to constrain the sampling process from structural distortion and modulate time step embedding features for better alignment with the input at the next time step. Third, to rapidly generalize the trained model to a new, unseen dose level with as few resources as possible, we devise a one-shot learning framework to make CoreDiff generalize faster and better using only one single LDCT image (un)paired with normal-dose CT (NDCT). Extensive experimental results on four datasets demonstrate that our CoreDiff outperforms competing methods in denoising and generalization performance, with clinically acceptable inference time. Source code is made available at https://github.com/qgao21/CoreDiff.
低剂量计算机断层扫描(CT)图像由于光子饥饿和电子噪声而受到噪声和伪影的影响。最近,一些工作试图使用扩散模型来解决以前基于深度学习的去噪模型所遇到的过度平滑和训练不稳定问题。然而,扩散模型由于涉及大量的采样步骤,因此推理时间较长。最近,冷扩散模型推广了经典的扩散模型,具有更大的灵活性。受冷扩散的启发,本文提出了一种新颖的基于上下文误差调制的广义扩散模型(CoreDiff)用于低剂量 CT(LDCT)去噪,称为 CoreDiff。首先,CoreDiff 使用 LDCT 图像来置换随机高斯噪声,并采用新颖的均值保持退化算子来模拟 CT 退化的物理过程,由于 LDCT 图像作为采样过程的起点,因此显著减少了采样步骤。其次,为了缓解采样过程中不完善的恢复算子引起的误差积累问题,我们提出了一种新颖的上下文误差调制恢复网络(CLEAR-Net),它可以利用上下文信息来约束采样过程,防止结构失真,并调制时间步嵌入特征,以更好地与下一个时间步的输入对齐。第三,为了尽可能少地使用资源,将训练好的模型快速推广到新的、未见过的剂量水平,我们设计了一种单样本学习框架,使 CoreDiff 使用仅一个 LDCT 图像(未配对)和正常剂量 CT(NDCT)更快、更好地进行泛化。在四个数据集上的广泛实验结果表明,我们的 CoreDiff 在去噪和泛化性能方面优于竞争方法,具有可接受的临床推理时间。源代码可在 https://github.com/qgao21/CoreDiff 上获得。