Feng Tao, Guo Yangyang, Huang Xiaoping, Qiao Yongliang
School of Internet, Anhui University, Hefei 230039, China.
National Engineering Research Center for Agro-Ecological Big Data Analysis & Application, Hefei 230039, China.
Animals (Basel). 2023 Aug 4;13(15):2521. doi: 10.3390/ani13152521.
Obtaining animal regions and the relative position relationship of animals in the scene is conducive to further studying animal habits, which is of great significance for smart animal farming. However, the complex breeding environment still makes detection difficult. To address the problems of poor target segmentation effects and the weak generalization ability of existing semantic segmentation models in complex scenes, a semantic segmentation model based on an improved DeepLabV3+ network (Imp-DeepLabV3+) was proposed. Firstly, the backbone network of the DeepLabV3+ model was replaced by MobileNetV2 to enhance the feature extraction capability of the model. Then, the layer-by-layer feature fusion method was adopted in the Decoder stage to integrate high-level semantic feature information with low-level high-resolution feature information at multi-scale to achieve more precise up-sampling operation. Finally, the SENet module was further introduced into the network to enhance information interaction after feature fusion and improve the segmentation precision of the model under complex datasets. The experimental results demonstrate that the Imp-DeepLabV3+ model achieved a high pixel accuracy () of 99.4%, a mean pixel accuracy () of 98.1%, and a mean intersection over union () of 96.8%. Compared to the original DeepLabV3+ model, the segmentation performance of the improved model significantly improved. Moreover, the overall segmentation performance of the Imp-DeepLabV3+ model surpassed that of other commonly used semantic segmentation models, such as Fully Convolutional Networks (FCNs), Lite Reduced Atrous Spatial Pyramid Pooling (LR-ASPP), and U-Net. Therefore, this study can be applied to the field of scene segmentation and is conducive to further analyzing individual information and promoting the development of intelligent animal farming.
获取场景中动物区域以及动物之间的相对位置关系,有助于进一步研究动物习性,这对智能畜牧养殖具有重要意义。然而,复杂的养殖环境仍然使得检测困难重重。为了解决现有语义分割模型在复杂场景中目标分割效果不佳和泛化能力较弱的问题,提出了一种基于改进的DeepLabV3+网络(Imp-DeepLabV3+)的语义分割模型。首先,将DeepLabV3+模型的骨干网络替换为MobileNetV2,以增强模型的特征提取能力。然后,在解码器阶段采用逐层特征融合方法,在多尺度下将高级语义特征信息与低级高分辨率特征信息进行整合,以实现更精确的上采样操作。最后,在网络中进一步引入SENet模块,以增强特征融合后的信息交互,并提高模型在复杂数据集下的分割精度。实验结果表明,Imp-DeepLabV3+模型实现了99.4%的高像素精度()、98.1%的平均像素精度()和96.8%的平均交并比()。与原始DeepLabV3+模型相比,改进模型的分割性能显著提高。此外,Imp-DeepLabV3+模型的整体分割性能超过了其他常用的语义分割模型,如全卷积网络(FCN)、轻量级空洞空间金字塔池化(LR-ASPP)和U-Net。因此,本研究可应用于场景分割领域,有助于进一步分析个体信息,推动智能畜牧养殖的发展。