IEEE Trans Neural Netw Learn Syst. 2022 Sep;33(9):4861-4875. doi: 10.1109/TNNLS.2021.3061630. Epub 2022 Aug 31.
It has been widely recognized that the efficient training of neural networks (NNs) is crucial to classification performance. While a series of gradient-based approaches have been extensively developed, they are criticized for the ease of trapping into local optima and sensitivity to hyperparameters. Due to the high robustness and wide applicability, evolutionary algorithms (EAs) have been regarded as a promising alternative for training NNs in recent years. However, EAs suffer from the curse of dimensionality and are inefficient in training deep NNs (DNNs). By inheriting the advantages of both the gradient-based approaches and EAs, this article proposes a gradient-guided evolutionary approach to train DNNs. The proposed approach suggests a novel genetic operator to optimize the weights in the search space, where the search direction is determined by the gradient of weights. Moreover, the network sparsity is considered in the proposed approach, which highly reduces the network complexity and alleviates overfitting. Experimental results on single-layer NNs, deep-layer NNs, recurrent NNs, and convolutional NNs (CNNs) demonstrate the effectiveness of the proposed approach. In short, this work not only introduces a novel approach for training DNNs but also enhances the performance of EAs in solving large-scale optimization problems.
人们已经广泛认识到,神经网络(NN)的有效训练对于分类性能至关重要。虽然已经开发了一系列基于梯度的方法,但它们很容易陷入局部最优解并且对超参数敏感,这受到了广泛的批评。由于具有高度的鲁棒性和广泛的适用性,进化算法(EAs)近年来被视为训练神经网络的一种有前途的替代方法。然而,EAs 受到维度诅咒的影响,在训练深度神经网络(DNN)时效率不高。本文通过继承基于梯度的方法和 EAs 的优点,提出了一种梯度引导的进化方法来训练 DNN。所提出的方法提出了一种新颖的遗传算子来优化搜索空间中的权重,其中搜索方向由权重的梯度决定。此外,所提出的方法考虑了网络稀疏性,这极大地降低了网络的复杂性并减轻了过拟合。在单层神经网络、深层神经网络、递归神经网络和卷积神经网络(CNN)上的实验结果证明了所提出方法的有效性。总之,这项工作不仅引入了一种训练 DNN 的新方法,而且提高了 EAs 在解决大规模优化问题方面的性能。