Shanxi Key Laboratory of Big Data Analysis and Parallel Computing, Taiyuan University of Science and Technology, Taiyuan, China.
Shanxi Key Laboratory of Big Data Analysis and Parallel Computing, Taiyuan University of Science and Technology, Taiyuan, China; State Key Laboratory for Novel Software Technology at Nanjing University, Nanjing University, Nanjing, China.
Comput Biol Med. 2024 Sep;180:108931. doi: 10.1016/j.compbiomed.2024.108931. Epub 2024 Jul 29.
Skin cancer images have hair occlusion problems, which greatly affects the accuracy of diagnosis and classification. Current dermoscopic hair removal methods use segmentation networks to locate hairs, and then uses repair networks to perform image repair. However, it is difficult to segment hair and capture the overall structure between hairs because of the hair being thin, unclear, and similar in color to the entire image. When conducting image restoration tasks, the only available images are those obstructed by hair, and there is no corresponding ground truth (supervised data) of the same scene without hair obstruction. In addition, the texture information and structural information used in existing repair methods are often insufficient, which leads to poor results in skin cancer image repair. To address these challenges, we propose the intersection-union dual-stream cross-attention Lova-SwinUnet (IUDC-LS). Firstly, we propose the Lova-SwinUnet module, which embeds Lovasz loss function into Swin-Unet, enabling the network to better capture features of various scales, thus obtaining better hair mask segmentation results. Secondly, we design the intersection-union (IU) module, which takes the mask results obtained in the previous step for pairwise intersection or union, and then overlays the results on the skin cancer image without hair to generate the labeled training data. This turns the unsupervised image repair task into the supervised one. Finally, we propose the dual-stream cross-attention (DC) module, which makes texture information and structure information interact with each other, and then uses cross-attention to make the network pay attention to the more important texture information and structure information in the fusion process of texture information and structure information, so as to improve the effect of image repair. The experimental results show that the PSNR index and SSIM index of the proposed method are increased by 5.4875 and 0.0401 compared with the other common methods. Experimental results unequivocally demonstrate the effectiveness of our approach, which serves as a potent tool for skin cancer detection, significantly surpassing the performance of comparable methods.
皮肤癌图像存在毛发遮挡问题,这极大地影响了诊断和分类的准确性。目前的皮肤镜去毛发方法使用分割网络定位毛发,然后使用修复网络进行图像修复。但是,由于毛发细、不清晰且与整个图像颜色相似,因此很难分割毛发并捕捉毛发之间的整体结构。在进行图像修复任务时,唯一可用的图像是被毛发遮挡的图像,而没有相应的无毛发遮挡的相同场景的真实图像(有监督数据)。此外,现有修复方法中使用的纹理信息和结构信息往往不足,这导致皮肤癌图像修复的效果较差。为了解决这些挑战,我们提出了交集并集双流交叉注意力 Lova-SwinUnet(IUDC-LS)。首先,我们提出了 Lova-SwinUnet 模块,将 Lovasz 损失函数嵌入到 Swin-Unet 中,使网络能够更好地捕获各种尺度的特征,从而获得更好的毛发掩模分割结果。其次,我们设计了交集并集(IU)模块,该模块对前一步获得的掩模结果进行两两交集或并集,然后将结果叠加在无毛发的皮肤癌图像上,生成带标签的训练数据。这将无监督的图像修复任务转化为有监督的任务。最后,我们提出了双流交叉注意力(DC)模块,该模块使纹理信息和结构信息相互作用,然后使用交叉注意力使网络在纹理信息和结构信息的融合过程中关注更重要的纹理信息和结构信息,从而提高图像修复的效果。实验结果表明,与其他常见方法相比,所提出方法的 PSNR 指标和 SSIM 指标分别提高了 5.4875 和 0.0401。实验结果明确证明了我们方法的有效性,它是皮肤癌检测的有力工具,显著优于可比方法的性能。