Wang Zhihe, Wang Huan, Du Hui, Chen Shiyin, Shi Xinxin
The School of Computer Science and Engineering, Northwest Normal University, Lanzhou 730070, China.
Math Biosci Eng. 2023 May 10;20(7):11875-11894. doi: 10.3934/mbe.2023528.
The density peak clustering algorithm (DPC) requires manual determination of cluster centers, and poor performance on complex datasets with varying densities or non-convexity. Hence, a novel density peak clustering algorithm is proposed for the automatic selection of clustering centers based on K-nearest neighbors (AKDPC). First, the AKDPC classifies samples according to their mutual K-nearest neighbor values into core and non-core points. Second, the AKDPC uses the average distance of K nearest neighbors of a sample as its density. The smaller the average distance is, the higher the density. Subsequently, it selects the highest density sample among all unclassified core points as a center of the new cluster, and the core points that satisfy the merging condition are added to the cluster until no core points satisfy the condition. Afterwards, the above steps are repeated to complete the clustering of all core points. Lastly, the AKDPC labels the unclassified non-core points similar to the nearest points that have been classified. In addition, to prove the validity of AKDPC, experiments on manual and real datasets are conducted. By comparing the AKDPC with classical clustering algorithms and excellent DPC-variants, this paper demonstrates that AKDPC presents higher accuracy.
密度峰值聚类算法(DPC)需要手动确定聚类中心,并且在密度变化或非凸性的复杂数据集上性能较差。因此,提出了一种基于K近邻的新型密度峰值聚类算法(AKDPC)用于自动选择聚类中心。首先,AKDPC根据样本的相互K近邻值将样本分类为核心点和非核心点。其次,AKDPC使用样本的K近邻平均距离作为其密度。平均距离越小,密度越高。随后,它在所有未分类的核心点中选择密度最高的样本作为新聚类的中心,并将满足合并条件的核心点添加到聚类中,直到没有核心点满足条件。之后,重复上述步骤以完成所有核心点的聚类。最后,AKDPC将未分类的非核心点标记为与已分类的最近点相似。此外,为了证明AKDPC的有效性,在人工和真实数据集上进行了实验。通过将AKDPC与经典聚类算法和优秀的DPC变体进行比较,本文表明AKDPC具有更高的准确性。