Zhang Ya-Jun, Liu Zhi-Qiang
Dept. of Comput. Sci. and Software Eng., Univ. of Melbourne, Vic.
IEEE Trans Neural Netw. 2002;13(2):369-80. doi: 10.1109/72.991422.
Clustering in the neural-network literature is generally based on the competitive learning paradigm. The paper addresses two major issues associated with conventional competitive learning, namely, sensitivity to initialization and difficulty in determining the number of prototypes. In general, selecting the appropriate number of prototypes is a difficult task, as we do not usually know the number of clusters in the input data a priori. It is therefore desirable to develop an algorithm that has no dependency on the initial prototype locations and is able to adaptively generate prototypes to fit the input data patterns. We present a new, more powerful competitive learning algorithm, self-splitting competitive learning (SSCL), that is able to find the natural number of clusters based on the one-prototype-take-one-cluster (OPTOC) paradigm and a self-splitting validity measure. It starts with a single prototype randomly initialized in the feature space and splits adaptively during the learning process until all clusters are found; each cluster is associated with a prototype at its center. We have conducted extensive experiments to demonstrate the effectiveness of the SSCL algorithm. The results show that SSCL has the desired ability for a variety of applications, including unsupervised classification, curve detection, and image segmentation.
神经网络文献中的聚类通常基于竞争学习范式。本文讨论了与传统竞争学习相关的两个主要问题,即对初始化的敏感性和确定原型数量的困难。一般来说,选择合适的原型数量是一项艰巨的任务,因为我们通常事先不知道输入数据中的聚类数量。因此,需要开发一种算法,该算法不依赖于初始原型位置,并且能够自适应地生成原型以适应输入数据模式。我们提出了一种新的、更强大的竞争学习算法,即自分裂竞争学习(SSCL),它能够基于单原型取单聚类(OPTOC)范式和自分裂有效性度量找到聚类的自然数量。它从在特征空间中随机初始化的单个原型开始,并在学习过程中自适应地分裂,直到找到所有聚类;每个聚类都与位于其中心的一个原型相关联。我们进行了广泛的实验来证明SSCL算法的有效性。结果表明,SSCL对于包括无监督分类、曲线检测和图像分割在内的各种应用具有所需的能力。