Department of Electrical and Electronic Engineering, Imperial College London, London SW7 2AZ, UK.
School of Electronic Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, China.
Sensors (Basel). 2022 Aug 18;22(16):6188. doi: 10.3390/s22166188.
RGB-D salient object detection (SOD) demonstrates its superiority in detecting in complex environments due to the additional depth information introduced in the data. Inevitably, an independent stream is introduced to extract features from depth images, leading to extra computation and parameters. This methodology sacrifices the model size to improve the detection accuracy which may impede the practical application of SOD problems. To tackle this dilemma, we propose a dynamic knowledge distillation (DKD) method, along with a lightweight structure, which significantly reduces the computational burden while maintaining validity. This method considers the factors of both teacher and student performance within the training stage and dynamically assigns the distillation weight instead of applying a fixed weight on the student model. We also investigate the issue of RGB-D early fusion strategy in distillation and propose a simple noise elimination method to mitigate the impact of distorted training data caused by low quality depth maps. Extensive experiments are conducted on five public datasets to demonstrate that our method can achieve competitive performance with a fast inference speed (136FPS) compared to 12 prior methods.
RGB-D 显著目标检测 (SOD) 由于在数据中引入了额外的深度信息,因此在检测复杂环境方面表现出优越性。不可避免地,引入了一个独立的流从深度图像中提取特征,导致额外的计算和参数。这种方法牺牲了模型大小来提高检测精度,这可能会阻碍 SOD 问题的实际应用。为了解决这个困境,我们提出了一种动态知识蒸馏 (DKD) 方法,以及一个轻量级的结构,在保持有效性的同时,显著降低了计算负担。该方法在训练阶段考虑了教师和学生表现的因素,并动态分配蒸馏权重,而不是在学生模型上应用固定权重。我们还研究了蒸馏中 RGB-D 早期融合策略的问题,并提出了一种简单的噪声消除方法来减轻低质量深度图导致的训练数据失真的影响。在五个公共数据集上进行了广泛的实验,结果表明,与 12 种先前的方法相比,我们的方法可以在快速推断速度(136FPS)下实现有竞争力的性能。