Chen Jianpin, Li Heng, Gao Qi, Liang Junling, Zhang Ruipeng, Yin Liping, Chai Xinyu
IEEE Trans Image Process. 2024;33:5593-5605. doi: 10.1109/TIP.2024.3470529. Epub 2024 Oct 9.
ConvNet-based object detection networks have achieved outstanding performance on clean images. However, many works have shown that these detectors perform poorly on corrupted images caused by noises, blurs, poor weather conditions and so on. With the development of security-sensitive applications, the detector's practicability has raised increasing concerns. Existing approaches improve detector robustness via extra operations (image restoration or training on extra labeled data) or by applying adversarial training at the expense of performance degradation on clean images. In this paper, we present Selective Adversarial Learning with Constraints (SALC) as a universal detector training approach to simultaneously improve the detector's precision and robustness. We first propose a unified formulation of adversarial samples for multitask adversarial learning, which significantly diversifies the obtained adversarial samples when integrated into the adversarial training of the detector. Next, we examine our findings on model bias against adversarial attacks of different strengths and differences in Batch Normalization (BN) statistics among clean images and different adversarial samples. On this basis, we propose a batch local comparison strategy with two BN branches to balance the detector's accuracy and robustness. Furthermore, to avoid performance degradation caused by overwhelming subtask losses, we leverage task-aware ratio thresholds to control the influence of learning in each subtask. The proposed approach can be applied to various detectors without any extra labeled data, inference time costs, or model parameters. Extensive experiments show that our SALC achieves state-of-the-art results on both clean benchmarks (Pascal VOC and MS-COCO) and corruption benchmarks (Pascal VOC-C and MS-COCO-C).
基于卷积神经网络(ConvNet)的目标检测网络在清晰图像上取得了出色的性能。然而,许多研究表明,这些检测器在由噪声、模糊、恶劣天气条件等导致的损坏图像上表现不佳。随着对安全敏感型应用的发展,检测器的实用性引起了越来越多的关注。现有方法通过额外操作(图像恢复或在额外标注数据上训练)或应用对抗训练来提高检测器的鲁棒性,但代价是在清晰图像上性能下降。在本文中,我们提出了带约束的选择性对抗学习(SALC)作为一种通用的检测器训练方法,以同时提高检测器的精度和鲁棒性。我们首先为多任务对抗学习提出了对抗样本的统一公式,当将其集成到检测器的对抗训练中时,能显著增加获得的对抗样本的多样性。接下来,我们研究了关于模型对不同强度对抗攻击的偏差以及清晰图像和不同对抗样本之间批归一化(BN)统计差异的发现。在此基础上,我们提出了一种带有两个BN分支的批局部比较策略,以平衡检测器的准确性和鲁棒性。此外,为了避免由压倒性的子任务损失导致的性能下降,我们利用任务感知比率阈值来控制每个子任务中学习的影响。所提出的方法可以应用于各种检测器,无需任何额外的标注数据、推理时间成本或模型参数。大量实验表明,我们的SALC在清晰基准(Pascal VOC和MS-COCO)和损坏基准(Pascal VOC-C和MS-COCO-C)上均取得了领先的结果。