IEEE Trans Cybern. 2022 Jun;52(6):5380-5393. doi: 10.1109/TCYB.2020.3031610. Epub 2022 Jun 16.
Due to its strong performance in handling uncertain and ambiguous data, the fuzzy k -nearest-neighbor method (FKNN) has realized substantial success in a wide variety of applications. However, its classification performance would be heavily deteriorated if the number k of nearest neighbors was unsuitably fixed for each testing sample. This study examines the feasibility of using only one fixed k value for FKNN on each testing sample. A novel FKNN-based classification method, namely, fuzzy KNN method with adaptive nearest neighbors (A-FKNN), is devised for learning a distinct optimal k value for each testing sample. In the training stage, after applying a sparse representation method on all training samples for reconstruction, A-FKNN learns the optimal k value for each training sample and builds a decision tree (namely, A-FKNN tree) from all training samples with new labels (the learned optimal k values instead of the original labels), in which each leaf node stores the corresponding optimal k value. In the testing stage, A-FKNN identifies the optimal k value for each testing sample by searching the A-FKNN tree and runs FKNN with the optimal k value for each testing sample. Moreover, a fast version of A-FKNN, namely, FA-FKNN, is designed by building the FA-FKNN decision tree, which stores the optimal k value with only a subset of training samples in each leaf node. Experimental results on 32 UCI datasets demonstrate that both A-FKNN and FA-FKNN outperform the compared methods in terms of classification accuracy, and FA-FKNN has a shorter running time.
由于在处理不确定和模糊数据方面的强大性能,模糊 k-最近邻方法(FKNN)在各种应用中取得了巨大的成功。然而,如果为每个测试样本固定的最近邻 k 值不合适,其分类性能将会严重恶化。本研究探讨了在每个测试样本上仅使用一个固定的 k 值进行 FKNN 的可行性。设计了一种基于模糊 k-最近邻的新分类方法,即具有自适应最近邻的模糊 k-最近邻方法(A-FKNN),用于为每个测试样本学习独特的最佳 k 值。在训练阶段,在对所有训练样本应用稀疏表示方法进行重构后,A-FKNN 为每个训练样本学习最佳 k 值,并使用新标签(学习到的最佳 k 值而不是原始标签)从所有训练样本构建决策树(即 A-FKNN 树),其中每个叶节点存储相应的最佳 k 值。在测试阶段,A-FKNN 通过搜索 A-FKNN 树为每个测试样本识别最佳 k 值,并为每个测试样本运行 FKNN 与最佳 k 值。此外,通过构建仅在每个叶节点中存储训练样本子集的最优 k 值的 FA-FKNN 决策树,设计了 FA-FKNN 的快速版本。在 32 个 UCI 数据集上的实验结果表明,A-FKNN 和 FA-FKNN 在分类准确性方面均优于比较方法,而 FA-FKNN 的运行时间更短。