Alioscha-Perez Mitchel, Oveneke Meshia Cedric, Sahli Hichem
IEEE Trans Neural Netw Learn Syst. 2020 May;31(5):1710-1723. doi: 10.1109/TNNLS.2019.2922123. Epub 2019 Jul 4.
In this paper, we present a novel strategy to combine a set of compact descriptors to leverage an associated recognition task. We formulate the problem from a multiple kernel learning (MKL) perspective and solve it following a stochastic variance reduced gradient (SVRG) approach to address its scalability, currently an open issue. MKL models are ideal candidates to jointly learn the optimal combination of features along with its associated predictor. However, they are unable to scale beyond a dozen thousand of samples due to high computational and memory requirements, which severely limits their applicability. We propose SVRG-MKL, an MKL solution with inherent scalability properties that can optimally combine multiple descriptors involving millions of samples. Our solution takes place directly in the primal to avoid Gram matrices computation and memory allocation, whereas the optimization is performed with a proposed algorithm of linear complexity and hence computationally efficient. Our proposition builds upon recent progress in SVRG with the distinction that each kernel is treated differently during optimization, which results in a faster convergence than applying off-the-shelf SVRG into MKL. Extensive experimental validation conducted on several benchmarking data sets confirms a higher accuracy and a significant speedup of our solution. Our technique can be extended to other MKL problems, including visual search and transfer learning, as well as other formulations, such as group-sensitive (GMKL) and localized MKL (LMKL) in convex settings.
在本文中,我们提出了一种新颖的策略,将一组紧凑描述符相结合以利用相关的识别任务。我们从多核学习(MKL)的角度来阐述该问题,并采用随机方差减少梯度(SVRG)方法来解决它,以应对其可扩展性这一当前的开放性问题。MKL模型是联合学习特征及其相关预测器的最优组合的理想候选者。然而,由于高计算和内存需求,它们无法处理超过一万个样本,这严重限制了它们的适用性。我们提出了SVRG - MKL,一种具有固有可扩展性的MKL解决方案,它可以最优地组合涉及数百万个样本的多个描述符。我们的解决方案直接在原始空间中进行,以避免计算Gram矩阵和内存分配,而优化是通过一种具有线性复杂度的算法进行的,因此计算效率高。我们的提议基于SVRG的最新进展,不同之处在于在优化过程中对每个核进行不同的处理,这导致比将现成的SVRG应用于MKL更快的收敛速度。在几个基准数据集上进行的广泛实验验证证实了我们的解决方案具有更高的准确性和显著的加速效果。我们的技术可以扩展到其他MKL问题,包括视觉搜索和迁移学习,以及其他形式,如凸设置下的组敏感(GMKL)和局部MKL(LMKL)。