School of Information Engineering, Pingdingshan University, Pingdingshan, China.
Faculty of Engineering, Built Environment and Information Technology, SEGi University, Kota Damansara, Malaysia.
PLoS One. 2024 Oct 3;19(10):e0309434. doi: 10.1371/journal.pone.0309434. eCollection 2024.
With the development of deep learning technology, convolutional neural networks have made great progress in the field of image segmentation. However, for complex scenes and multi-scale target images, the existing technologies are still unable to achieve effective image segmentation. In view of this, an image segmentation model based on residual connection and feature fusion is proposed. The model makes comprehensive use of the deep feature extraction ability of residual connections and the multi-scale feature integration ability of feature fusion. In order to solve the problem of background complexity and information loss in traditional image segmentation, experiments were carried out on two publicly available data sets. The results showed that in the ISPRS Vaihingen dataset and the Caltech UCSD Birds200 dataset, when the model completed the 56th and 84th iterations, respectively, the average accuracy of FRes-MFDNN was the highest, which was 97.89% and 98.24%, respectively. In the ISPRS Vaihingen dataset and the Caltech UCSD Birds200 dataset, when the system model ran to 0.20s and 0.26s, the F1 value of the FRes-MFDNN method was the largest, and the F1 value approached 100% infinitely. The FRes-MFDNN segmented four images in the ISPRS Vaihingen dataset, and the segmentation accuracy of images 1, 2, 3 and 4 were 91.44%, 92.12%, 94.02% and 91.41%, respectively. In practical applications, the MSRF-Net method, LBN-AA-SPN method, ARG-Otsu method, and FRes-MFDNN were used to segment unlabeled bird images. The results showed that the FRes-MFDNN was more complete in details, and the overall effect was significantly better than the other three models. Meanwhile, in ordinary scene images, although there was a certain degree of noise and occlusion, the model still accurately recognized and segmented the main bird images. The results show that compared with the traditional model, after FRes-MFDNN segmentation, the completeness, detail, and spatial continuity of pixels have been significantly improved, making it more suitable for complex scenes.
随着深度学习技术的发展,卷积神经网络在图像分割领域取得了很大的进展。然而,对于复杂场景和多尺度目标图像,现有的技术仍然无法实现有效的图像分割。针对这一问题,提出了一种基于残差连接和特征融合的图像分割模型。该模型充分利用了残差连接的深度特征提取能力和特征融合的多尺度特征融合能力。为了解决传统图像分割中背景复杂和信息丢失的问题,在两个公开可用的数据集上进行了实验。结果表明,在 ISPRS Vaihingen 数据集和 Caltech UCSD Birds200 数据集上,当模型分别完成 56 次和 84 次迭代时,FRes-MFDNN 的平均准确率最高,分别为 97.89%和 98.24%。在 ISPRS Vaihingen 数据集和 Caltech UCSD Birds200 数据集上,当系统模型运行到 0.20s 和 0.26s 时,FRes-MFDNN 方法的 F1 值最大,并且无限接近 100%。FRes-MFDNN 对 ISPRS Vaihingen 数据集中的四张图像进行了分割,图像 1、2、3 和 4 的分割准确率分别为 91.44%、92.12%、94.02%和 91.41%。在实际应用中,使用 MSRF-Net 方法、LBN-AA-SPN 方法、ARG-Otsu 方法和 FRes-MFDNN 方法对未标记的鸟类图像进行分割。结果表明,FRes-MFDNN 在细节上更加完整,整体效果明显优于其他三个模型。同时,在普通场景图像中,虽然存在一定程度的噪声和遮挡,但模型仍然能够准确识别和分割主要的鸟类图像。结果表明,与传统模型相比,经过 FRes-MFDNN 分割后,像素的完整性、细节和空间连续性都得到了显著提高,使其更适用于复杂场景。