Jin Xiao-Bo, Zhang Xu-Yao, Huang Kaizhu, Geng Guang-Gang
IEEE Trans Neural Netw Learn Syst. 2019 May;30(5):1360-1369. doi: 10.1109/TNNLS.2018.2868835. Epub 2018 Sep 27.
Conjugate gradient (CG) methods are a class of important methods for solving linear equations and nonlinear optimization problems. In this paper, we propose a new stochastic CG algorithm with variance reduction and we prove its linear convergence with the Fletcher and Reeves method for strongly convex and smooth functions. We experimentally demonstrate that the CG with variance reduction algorithm converges faster than its counterparts for four learning models, which may be convex, nonconvex or nonsmooth. In addition, its area under the curve performance on six large-scale data sets is comparable to that of the LIBLINEAR solver for the L2 -regularized L2 -loss but with a significant improvement in computational efficiency.CGVR algorithm is available on github: https://github.com/xbjin/cgvr.
共轭梯度(CG)方法是一类求解线性方程和非线性优化问题的重要方法。在本文中,我们提出了一种新的具有方差缩减的随机共轭梯度算法,并使用Fletcher-Reeves方法证明了其对于强凸且光滑函数的线性收敛性。我们通过实验证明,对于四种可能是凸、非凸或非光滑的学习模型,具有方差缩减的共轭梯度算法比其他同类算法收敛得更快。此外,在六个大规模数据集上,其曲线下面积性能与用于L2正则化L2损失的LIBLINEAR求解器相当,但计算效率有显著提高。CGVR算法可在github上获取:https://github.com/xbjin/cgvr 。