Luo Sheng, Ma Yiming, Jiang Feng, Wang Hongying, Tong Qin, Wang Liangju
College of Engineering, China Agricultural University, Beijing 100083, China.
College of Water Resources and Civil Engineering, China Agricultural University, Beijing 100083, China.
Animals (Basel). 2023 Jun 2;13(11):1861. doi: 10.3390/ani13111861.
In large-scale laying hen farming, timely detection of dead chickens helps prevent cross-infection, disease transmission, and economic loss. Dead chicken detection is still performed manually and is one of the major labor costs on commercial farms. This study proposed a new method for dead chicken detection using multi-source images and deep learning and evaluated the detection performance with different source images. We first introduced a pixel-level image registration method that used depth information to project the near-infrared (NIR) and depth image into the coordinate of the thermal infrared (TIR) image, resulting in registered images. Then, the registered single-source (TIR, NIR, depth), dual-source (TIR-NIR, TIR-depth, NIR-depth), and multi-source (TIR-NIR-depth) images were separately used to train dead chicken detecting models with object detection networks, including YOLOv8n, Deformable DETR, Cascade R-CNN, and TOOD. The results showed that, at an IoU (Intersection over Union) threshold of 0.5, the performance of these models was not entirely the same. Among them, the model using the NIR-depth image and Deformable DETR achieved the best performance, with an average precision (AP) of 99.7% (IoU = 0.5) and a recall of 99.0% (IoU = 0.5). While the IoU threshold increased, we found the following: The model with the NIR image achieved the best performance among models with single-source images, with an AP of 74.4% (IoU = 0.5:0.95) in Deformable DETR. The performance with dual-source images was higher than that with single-source images. The model with the TIR-NIR or NIR-depth image outperformed the model with the TIR-depth image, achieving an AP of 76.3% (IoU = 0.5:0.95) and 75.9% (IoU = 0.5:0.95) in Deformable DETR, respectively. The model with the multi-source image also achieved higher performance than that with single-source images. However, there was no significant improvement compared to the model with the TIR-NIR or NIR-depth image, and the AP of the model with multi-source image was 76.7% (IoU = 0.5:0.95) in Deformable DETR. By analyzing the detection performance with different source images, this study provided a reference for selecting and using multi-source images for detecting dead laying hens on commercial farms.
在大规模蛋鸡养殖中,及时检测死鸡有助于预防交叉感染、疾病传播和经济损失。死鸡检测目前仍需人工进行,是商业养殖场的主要劳动力成本之一。本研究提出了一种利用多源图像和深度学习进行死鸡检测的新方法,并对不同源图像的检测性能进行了评估。我们首先引入了一种像素级图像配准方法,该方法利用深度信息将近红外(NIR)图像和深度图像投影到热红外(TIR)图像的坐标中,从而得到配准图像。然后,分别使用配准后的单源(TIR、NIR、深度)、双源(TIR-NIR、TIR-深度、NIR-深度)和多源(TIR-NIR-深度)图像,通过目标检测网络(包括YOLOv8n、可变形DETR、级联R-CNN和TOOD)训练死鸡检测模型。结果表明,在交并比(IoU)阈值为0.5时,这些模型的性能并不完全相同。其中,使用NIR-深度图像和可变形DETR的模型性能最佳,平均精度(AP)为99.7%(IoU = 0.5),召回率为99.0%(IoU = 0.5)。当IoU阈值增加时,我们发现:在单源图像模型中,使用NIR图像的模型性能最佳,在可变形DETR中,AP为74.4%(IoU = 0.5:0.95)。双源图像的性能高于单源图像。使用TIR-NIR或NIR-深度图像的模型优于使用TIR-深度图像的模型,在可变形DETR中,AP分别为76.3%(IoU = 0.5:0.95)和75.9%(IoU = 0.5:0.95)。多源图像模型的性能也高于单源图像模型。然而,与使用TIR-NIR或NIR-深度图像的模型相比,没有显著提高,在可变形DETR中,多源图像模型的AP为76.7%(IoU = 0.5:0.95)。通过分析不同源图像的检测性能,本研究为商业养殖场选择和使用多源图像检测死蛋鸡提供了参考。