Zhou Sihang, Liu Xinwang, Li Miaomiao, Zhu En, Liu Li, Zhang Changwang, Yin Jianping
IEEE Trans Neural Netw Learn Syst. 2020 Apr;31(4):1351-1362. doi: 10.1109/TNNLS.2019.2919900. Epub 2019 Jun 28.
Multiple kernel clustering (MKC) has been intensively studied during the last few decades. Even though they demonstrate promising clustering performance in various applications, existing MKC algorithms do not sufficiently consider the intrinsic neighborhood structure among base kernels, which could adversely affect the clustering performance. In this paper, we propose a simple yet effective neighbor-kernel-based MKC algorithm to address this issue. Specifically, we first define a neighbor kernel, which can be utilized to preserve the block diagonal structure and strengthen the robustness against noise and outliers among base kernels. After that, we linearly combine these base neighbor kernels to extract a consensus affinity matrix through an exact-rank-constrained subspace segmentation. The naturally possessed block diagonal structure of neighbor kernels better serves the subsequent subspace segmentation, and in turn, the extracted shared structure is further refined through subspace segmentation based on the combined neighbor kernels. In this manner, the above two learning processes can be seamlessly coupled and negotiate with each other to achieve better clustering. Furthermore, we carefully design an efficient iterative optimization algorithm with proven convergence to address the resultant optimization problem. As a by-product, we reveal an interesting insight into the exact-rank constraint in ridge regression by careful theoretical analysis: it back-projects the solution of the unconstrained counterpart to its principal components. Comprehensive experiments have been conducted on several benchmark data sets, and the results demonstrate the effectiveness of the proposed algorithm.
在过去几十年中,多核聚类(MKC)受到了广泛研究。尽管它们在各种应用中展现出了有前景的聚类性能,但现有的MKC算法没有充分考虑基核之间的内在邻域结构,这可能会对聚类性能产生不利影响。在本文中,我们提出了一种简单而有效的基于邻域核的MKC算法来解决这个问题。具体来说,我们首先定义一个邻域核,它可用于保留块对角结构并增强对基核中噪声和离群值的鲁棒性。之后,我们通过精确秩约束子空间分割将这些基邻域核进行线性组合,以提取一个共识亲和矩阵。邻域核自然具有的块对角结构更好地服务于后续的子空间分割,反过来,基于组合邻域核的子空间分割进一步细化提取的共享结构。通过这种方式,上述两个学习过程可以无缝耦合并相互协调以实现更好的聚类。此外,我们精心设计了一种具有收敛性证明的高效迭代优化算法来解决由此产生的优化问题。作为一个附带成果,我们通过仔细的理论分析揭示了关于岭回归中精确秩约束的一个有趣见解:它将无约束对应问题的解反向投影到其主成分上。我们在几个基准数据集上进行了全面的实验,结果证明了所提算法的有效性。