Department of Neuroscience, Baylor College of Medicine, Houston, Texas, United States of America.
University of Tübingen, Germany.
PLoS Comput Biol. 2023 Mar 27;19(3):e1010932. doi: 10.1371/journal.pcbi.1010932. eCollection 2023 Mar.
Machine learning models have difficulty generalizing to data outside of the distribution they were trained on. In particular, vision models are usually vulnerable to adversarial attacks or common corruptions, to which the human visual system is robust. Recent studies have found that regularizing machine learning models to favor brain-like representations can improve model robustness, but it is unclear why. We hypothesize that the increased model robustness is partly due to the low spatial frequency preference inherited from the neural representation. We tested this simple hypothesis with several frequency-oriented analyses, including the design and use of hybrid images to probe model frequency sensitivity directly. We also examined many other publicly available robust models that were trained on adversarial images or with data augmentation, and found that all these robust models showed a greater preference to low spatial frequency information. We show that preprocessing by blurring can serve as a defense mechanism against both adversarial attacks and common corruptions, further confirming our hypothesis and demonstrating the utility of low spatial frequency information in robust object recognition.
机器学习模型很难泛化到其训练数据之外的分布。特别是,视觉模型通常容易受到对抗攻击或常见的损坏,而人类的视觉系统则具有很强的鲁棒性。最近的研究发现,对机器学习模型进行正则化以偏向于类脑表示可以提高模型的鲁棒性,但原因尚不清楚。我们假设,模型鲁棒性的提高部分归因于从神经表示中继承的低空间频率偏好。我们通过几种面向频率的分析来测试这个简单的假设,包括设计和使用混合图像来直接探测模型的频率敏感性。我们还研究了许多其他基于对抗图像或数据增强训练的可用的鲁棒模型,发现所有这些鲁棒模型都表现出对低空间频率信息的更大偏好。我们表明,通过模糊处理进行预处理可以作为对抗攻击和常见损坏的防御机制,进一步证实了我们的假设,并展示了低空间频率信息在鲁棒目标识别中的应用。