College of Information Technology, Shanghai Ocean University, Shanghai 201306, China.
Guangdong Feida Transportation Engineering Co., Ltd., Guangzhou 510663, China.
Sensors (Basel). 2022 Sep 22;22(19):7204. doi: 10.3390/s22197204.
Aiming at the common problems, such as noise pollution, low contrast, and color distortion in underwater images, and the characteristics of holothurian recognition, such as morphological ambiguity, high similarity with the background, and coexistence of special ecological scenes, this paper proposes an underwater holothurian target-detection algorithm (FA-CenterNet), based on improved CenterNet and scene feature fusion. First, to reduce the model's occupancy of embedded device resources, we use EfficientNet-B3 as the backbone network to reduce the model's Params and FLOPs. At the same time, EfficientNet-B3 increases the depth and width of the model, which improves the accuracy of the model. Then, we design an effective FPT (feature pyramid transformer) combination module to fully focus and mine the information on holothurian ecological scenarios of different scales and spaces (e.g., holothurian spines, reefs, and waterweeds are often present in the same scenario as holothurians). The co-existing scene information can be used as auxiliary features to detect holothurians, which can improve the detection ability of fuzzy and small-sized holothurians. Finally, we add the AFF module to realize the deep fusion of the shallow-detail and high-level semantic features of holothurians. The results show that the method presented in this paper yields better results on the 2020 CURPC underwater target-detection image dataset with an AP50 of 83.43%, Params of 15.90 M, and FLOPs of 25.12 G compared to other methods. In the underwater holothurian-detection task, this method improves the accuracy of detecting holothurians with fuzzy features, a small size, and dense scene. It also achieves a good balance between detection accuracy, Params, and FLOPs, and is suitable for underwater holothurian detection in most situations.
针对水下图像常见的噪声污染、对比度低、颜色失真以及海参识别的形态模糊、与背景相似度高、特殊生态场景共存等问题,本文提出了一种基于改进 CenterNet 和场景特征融合的水下海参目标检测算法(FA-CenterNet)。首先,为了减少模型对嵌入式设备资源的占用,我们使用 EfficientNet-B3 作为骨干网络,减少模型的参数量和 FLOPs。同时,EfficientNet-B3 增加了模型的深度和宽度,提高了模型的精度。然后,我们设计了一个有效的 FPT(特征金字塔变换)组合模块,充分关注和挖掘不同尺度和空间的海参生态场景信息(例如,海参的刺、珊瑚和水草经常与海参同时存在于同一场景中)。共存场景信息可作为辅助特征用于海参检测,可提高对模糊和小尺寸海参的检测能力。最后,我们添加了 AFF 模块,实现了海参浅层细节和高层语义特征的深度融合。实验结果表明,与其他方法相比,本文方法在 2020 CURPC 水下目标检测图像数据集上取得了更好的效果,AP50 为 83.43%,参数量为 15.90M,FLOPs 为 25.12G。在水下海参检测任务中,该方法提高了对具有模糊特征、小尺寸和密集场景的海参的检测精度,在检测精度、参数量和 FLOPs 之间取得了良好的平衡,适用于大多数水下海参检测场景。