Xu Yong-Li, Li Xiao-Xing, Chen Di-Rong, Li Han-Xiong
IEEE Trans Neural Netw Learn Syst. 2018 Nov;29(11):5408-5418. doi: 10.1109/TNNLS.2018.2802469. Epub 2018 Mar 2.
This paper considers a least square regularized regression algorithm for multi-task learning in a union of reproducing kernel Hilbert spaces (RKHSs) with Gaussian kernels. It is assumed that the optimal prediction function of the target task and those of related tasks are in an RKHS with the same but with unknown Gaussian kernel width. The samples for related tasks are used to select the Gaussian kernel width, and the sample for the target task is used to obtain the prediction function in the RKHS with this selected width. With an error decomposition result, a fast learning rate is obtained for the target task. The key step is to estimate the sample errors of related tasks in the union of RKHSs with Gaussian kernels. The utility of this algorithm is illustrated with one simulated data set and four real data sets. The experiment results illustrate that the underlying algorithm can result in significant improvements in prediction error when few samples of the target task and more samples of related tasks are available.
本文考虑了一种用于在具有高斯核的再生核希尔伯特空间(RKHSs)联合中进行多任务学习的最小二乘正则化回归算法。假设目标任务的最优预测函数以及相关任务的最优预测函数位于具有相同但未知高斯核宽度的RKHS中。相关任务的样本用于选择高斯核宽度,目标任务的样本用于在具有所选宽度的RKHS中获得预测函数。通过误差分解结果,为目标任务获得了快速学习率。关键步骤是估计具有高斯核的RKHS联合中相关任务的样本误差。用一个模拟数据集和四个真实数据集说明了该算法的效用。实验结果表明,当目标任务的样本较少而相关任务的样本较多时,底层算法可以显著提高预测误差。