Wan Weitao, Yu Cheng, Chen Jiansheng, Wu Tong, Zhong Yuanyi, Yang Ming-Hsuan
IEEE Trans Pattern Anal Mach Intell. 2023 Feb;45(2):2430-2444. doi: 10.1109/TPAMI.2022.3166879. Epub 2023 Jan 6.
The softmax cross-entropy loss function has been widely used to train deep models for various tasks. In this work, we propose a Gaussian mixture (GM) loss function for deep neural networks for visual classification. Unlike the softmax cross-entropy loss, our method explicitly shapes the deep feature space towards a Gaussian Mixture distribution. With a classification margin and a likelihood regularization, the GM loss facilitates both high classification performance and accurate modeling of the feature distribution. The GM loss can be readily used to distinguish the adversarial examples based on the discrepancy between feature distributions of clean and adversarial examples. Furthermore, theoretical analysis shows that a symmetric feature space can be achieved by using the GM loss, which enables the models to perform robustly against adversarial attacks. The proposed model can be implemented easily and efficiently without introducing more trainable parameters. Extensive evaluations demonstrate that the method with the GM loss performs favorably on image classification, face recognition, and detection as well as recognition of adversarial examples generated by various attacks.
softmax交叉熵损失函数已被广泛用于训练针对各种任务的深度模型。在这项工作中,我们为用于视觉分类的深度神经网络提出了一种高斯混合(GM)损失函数。与softmax交叉熵损失不同,我们的方法将深度特征空间明确地塑造为高斯混合分布。通过分类边界和似然正则化,GM损失既有助于实现高分类性能,又有助于对特征分布进行准确建模。GM损失可根据干净样本和对抗样本的特征分布差异,轻松用于区分对抗样本。此外,理论分析表明,使用GM损失可以实现对称特征空间,这使模型能够对对抗攻击表现出鲁棒性。所提出的模型可以轻松高效地实现,而无需引入更多可训练参数。广泛的评估表明,采用GM损失的方法在图像分类、人脸识别、检测以及识别由各种攻击生成的对抗样本方面表现出色。