School of Artificial Intelligence, Chongqing University of Technology, Chongqing, China.
School of Artificial Intelligence, Chongqing University of Technology, Chongqing, China; School of Computing and College of Design and Engineering, National University of Singapore, Singapore.
Comput Biol Med. 2024 Jul;177:108625. doi: 10.1016/j.compbiomed.2024.108625. Epub 2024 May 21.
Liver segmentation is a fundamental prerequisite for the diagnosis and surgical planning of hepatocellular carcinoma. Traditionally, the liver contour is drawn manually by radiologists using a slice-by-slice method. However, this process is time-consuming and error-prone, depending on the radiologist's experience. In this paper, we propose a new end-to-end automatic liver segmentation framework, named ResTransUNet, which exploits the transformer's ability to capture global context for remote interactions and spatial relationships, as well as the excellent performance of the original U-Net architecture. The main contribution of this paper lies in proposing a novel fusion network that combines Unet and Transformer architectures. In the encoding structure, a dual-path approach is utilized, where features are extracted separately using both convolutional neural networks (CNNs) and Transformer networks. Additionally, an effective feature enhancement unit is designed to transfer the global features extracted by the Transformer network to the CNN for feature enhancement. This model aims to address the drawbacks of traditional Unet-based methods, such as feature loss during encoding and poor capture of global features. Moreover, it avoids the disadvantages of pure Transformer models, which suffer from large parameter sizes and high computational complexity. The experimental results on the LiTS2017 dataset demonstrate remarkable performance for our proposed model, with Dice coefficients, volumetric overlap error (VOE), and relative volume difference (RVD) values for liver segmentation reaching 0.9535, 0.0804, and -0.0007, respectively. Furthermore, to further validate the model's generalization capability, we conducted tests on the 3Dircadb, Chaos, and Sliver07 datasets. The experimental results demonstrate that the proposed method outperforms other closely related models with higher liver segmentation accuracy. In addition, significant improvements can be achieved by applying our method when handling liver segmentation with small and discontinuous liver regions, as well as blurred liver boundaries. The code is available at the website: https://github.com/Jouiry/ResTransUNet.
肝脏分割是肝细胞癌诊断和手术规划的基本前提。传统上,放射科医生使用逐层方法手动绘制肝脏轮廓。然而,这个过程耗时且容易出错,取决于放射科医生的经验。在本文中,我们提出了一种新的端到端自动肝脏分割框架,名为 ResTransUNet,它利用了变压器捕捉远程交互和空间关系的全局上下文的能力,以及原始 U-Net 架构的优异性能。本文的主要贡献在于提出了一种新的融合网络,该网络结合了 U-Net 和 Transformer 架构。在编码结构中,采用了双路径方法,分别使用卷积神经网络 (CNN) 和 Transformer 网络提取特征。此外,设计了一种有效的特征增强单元,将 Transformer 网络提取的全局特征传递到 CNN 进行特征增强。该模型旨在解决传统基于 U-Net 的方法的缺点,例如编码过程中的特征丢失和全局特征的捕获不佳。此外,它避免了纯 Transformer 模型的缺点,这些模型存在参数规模大和计算复杂度高的问题。在 LiTS2017 数据集上的实验结果表明,我们提出的模型具有出色的性能,肝脏分割的 Dice 系数、体积重叠误差 (VOE) 和相对体积差异 (RVD) 值分别达到 0.9535、0.0804 和-0.0007。此外,为了进一步验证模型的泛化能力,我们在 3Dircadb、Chaos 和 Sliver07 数据集上进行了测试。实验结果表明,该方法在处理小而不连续的肝脏区域和模糊的肝脏边界的肝脏分割时,比其他密切相关的模型具有更高的肝脏分割精度。此外,通过应用我们的方法,可以显著提高肝脏分割的精度。该代码可在网站上获得:https://github.com/Jouiry/ResTransUNet。