Department of Computer Science, University of Management and Technology, Lahore, Pakistan.
Department of Computer Engineering, Bahria University Islamabad, Pakistan.
PLoS One. 2023 Apr 26;18(4):e0284992. doi: 10.1371/journal.pone.0284992. eCollection 2023.
Regular monitoring of the number of various fish species in a variety of habitats is essential for marine conservation efforts and marine biology research. To address the shortcomings of existing manual underwater video fish sampling methods, a plethora of computer-based techniques are proposed. However, there is no perfect approach for the automated identification and categorizing of fish species. This is primarily due to the difficulties inherent in capturing underwater videos, such as ambient changes in luminance, fish camouflage, dynamic environments, watercolor, poor resolution, shape variation of moving fish, and tiny differences between certain fish species. This study has proposed a novel Fish Detection Network (FD_Net) for the detection of nine different types of fish species using a camera-captured image that is based on the improved YOLOv7 algorithm by exchanging Darknet53 for MobileNetv3 and depthwise separable convolution for 3 x 3 filter size in the augmented feature extraction network bottleneck attention module (BNAM). The mean average precision (mAP) is 14.29% higher than it was in the initial version of YOLOv7. The network that is utilized in the method for the extraction of features is an improved version of DenseNet-169, and the loss function is an Arcface Loss. Widening the receptive field and improving the capability of feature extraction are achieved by incorporating dilated convolution into the dense block, removing the max-pooling layer from the trunk, and incorporating the BNAM into the dense block of the DenseNet-169 neural network. The results of several experiments comparisons and ablation experiments demonstrate that our proposed FD_Net has a higher detection mAP than YOLOv3, YOLOv3-TL, YOLOv3-BL, YOLOv4, YOLOv5, Faster-RCNN, and the most recent YOLOv7 model, and is more accurate for target fish species detection tasks in complex environments.
定期监测各种栖息地中各种鱼类的数量,对于海洋保护工作和海洋生物学研究至关重要。为了解决现有人工水下视频鱼类采样方法的不足,提出了许多基于计算机的技术。然而,对于鱼类物种的自动识别和分类,目前还没有一种完美的方法。这主要是由于水下视频拍摄存在固有困难,例如环境亮度变化、鱼类伪装、动态环境、水色、低分辨率、游动鱼类的形状变化以及某些鱼类之间的微小差异。本研究提出了一种新的 Fish Detection Network(FD_Net),用于使用相机拍摄的图像检测九种不同类型的鱼类,该图像基于改进的 YOLOv7 算法,通过在增强特征提取网络瓶颈注意力模块(BNAM)中用 MobileNetv3 替换 Darknet53 并将 3x3 滤波器大小的深度可分离卷积替换为 3x3 滤波器大小的深度可分离卷积,从而实现了九种不同类型的鱼类检测。与最初的 YOLOv7 版本相比,平均精度(mAP)提高了 14.29%。所使用的特征提取方法的网络是改进的 DenseNet-169 版本,损失函数是 Arcface Loss。通过在密集块中加入扩张卷积、从主干中移除最大池化层以及在 DenseNet-169 神经网络的密集块中加入 BNAM,拓宽了接收场并提高了特征提取能力。通过几个实验比较和消融实验的结果表明,我们提出的 FD_Net 比 YOLOv3、YOLOv3-TL、YOLOv3-BL、YOLOv4、YOLOv5、Faster-RCNN 和最新的 YOLOv7 模型具有更高的检测 mAP,并且在复杂环境中的目标鱼类检测任务中更加准确。