School of Advanced Technologoy, Xi'an Jiaotong-Liverpool University, Suzhou, 215123, China.
Department of Computer Science, Zhejiang University, Zhejiang, China.
Neural Netw. 2021 Aug;140:282-293. doi: 10.1016/j.neunet.2021.03.031. Epub 2021 Mar 26.
We propose a new regularization method for deep learning based on the manifold adversarial training (MAT). Unlike previous regularization and adversarial training methods, MAT further considers the local manifold of latent representations. Specifically, MAT manages to build an adversarial framework based on how the worst perturbation could affect the statistical manifold in the latent space rather than the output space. Particularly, a latent feature space with the Gaussian Mixture Model (GMM) is first derived in a deep neural network. We then define the smoothness by the largest variation of Gaussian mixtures when a local perturbation is given around the input data point. On one hand, the perturbations are added in the way that would rough the statistical manifold of the latent space the worst. On the other hand, the model is trained to promote the manifold smoothness the most in the latent space. Importantly, since the latent space is more informative than the output space, the proposed MAT can learn a more robust and compact data representation, leading to further performance improvement. The proposed MAT is important in that it can be considered as a superset of one recently-proposed discriminative feature learning approach called center loss. We conduct a series of experiments in both supervised and semi-supervised learning on four benchmark data sets, showing that the proposed MAT can achieve remarkable performance, much better than those of the state-of-the-art approaches. In addition, we present a series of visualization which could generate further understanding or explanation on adversarial examples.
我们提出了一种新的基于流形对抗训练(MAT)的深度学习正则化方法。与之前的正则化和对抗训练方法不同,MAT 进一步考虑了潜在表示的局部流形。具体来说,MAT 设法基于最恶劣的扰动如何影响潜在空间中的统计流形而不是输出空间来构建对抗框架。特别是,首先在深度神经网络中推导出具有高斯混合模型(GMM)的潜在特征空间。然后,我们通过在输入数据点周围给定局部扰动时高斯混合的最大变化来定义平滑度。一方面,通过添加扰动的方式使潜在空间的统计流形变得最粗糙。另一方面,对模型进行训练以在潜在空间中最大程度地促进流形平滑度。重要的是,由于潜在空间比输出空间更具信息量,因此所提出的 MAT 可以学习更稳健和紧凑的数据表示,从而进一步提高性能。所提出的 MAT 很重要,因为它可以被视为最近提出的一种称为中心损失的有区别的特征学习方法的超集。我们在四个基准数据集上进行了一系列监督和半监督学习实验,结果表明所提出的 MAT 可以实现显著的性能,远远优于最新方法。此外,我们还展示了一系列可视化结果,可以进一步理解或解释对抗示例。