Pang Yanwei, Zhou Bo, Nie Feiping
IEEE Trans Neural Netw Learn Syst. 2019 Sep;30(9):2779-2793. doi: 10.1109/TNNLS.2018.2886317. Epub 2019 Jan 10.
Explicitly or implicitly, most dimensionality reduction methods need to determine which samples are neighbors and the similarities between the neighbors in the original high-dimensional space. The projection matrix is then learnt on the assumption that the neighborhood information, e.g., the similarities, are known and fixed prior to learning. However, it is difficult to precisely measure the intrinsic similarities of samples in high-dimensional space because of the curse of dimensionality. Consequently, the neighbors selected according to such similarities and the projection matrix obtained according to such similarities and the corresponding neighbors might not be optimal in the sense of classification and generalization. To overcome this drawback, in this paper, we propose to let the similarities and neighbors be variables and model these in a low-dimensional space. Both the optimal similarity and projection matrix are obtained by minimizing a unified objective function. Nonnegative and sum-to-one constraints on the similarity are adopted. Instead of empirically setting the regularization parameter, we treat it as a variable to be optimized. It is interesting that the optimal regularization parameter is adaptive to the neighbors in a low-dimensional space and has an intuitive meaning. Experimental results on the YALE B, COIL-100, and MNIST data sets demonstrate the effectiveness of the proposed method.
大多数降维方法都直接或间接地需要确定哪些样本是邻居以及原始高维空间中邻居之间的相似性。然后,在假设邻域信息(例如相似性)在学习之前已知且固定的情况下学习投影矩阵。然而,由于维度诅咒,很难精确测量高维空间中样本的内在相似性。因此,根据这种相似性选择的邻居以及根据这种相似性和相应邻居获得的投影矩阵在分类和泛化意义上可能不是最优的。为了克服这一缺点,在本文中,我们建议将相似性和邻居设为变量,并在低维空间中对其进行建模。通过最小化一个统一的目标函数来获得最优相似性和投影矩阵。对相似性采用非负和归一化约束。我们不是凭经验设置正则化参数,而是将其视为一个待优化的变量。有趣的是,最优正则化参数适应低维空间中的邻居并且具有直观的意义。在YALE B、COIL - 100和MNIST数据集上的实验结果证明了所提方法的有效性。