Tao Dacheng, Li Xuelong, Wu Xindong, Maybank Stephen J
School of Computer Engineering, Nanyang Technological University, Singapore.
IEEE Trans Pattern Anal Mach Intell. 2009 Feb;31(2):260-74. doi: 10.1109/TPAMI.2008.70.
Subspace selection approaches are powerful tools in pattern classification and data visualization. One of the most important subspace approaches is the linear dimensionality reduction step in the Fisher's linear discriminant analysis (FLDA), which has been successfully employed in many fields such as biometrics, bioinformatics, and multimedia information management. However, the linear dimensionality reduction step in FLDA has a critical drawback: for a classification task with c classes, if the dimension of the projected subspace is strictly lower than c - 1, the projection to a subspace tends to merge those classes, which are close together in the original feature space. If separate classes are sampled from Gaussian distributions, all with identical covariance matrices, then the linear dimensionality reduction step in FLDA maximizes the mean value of the Kullback-Leibler (KL) divergences between different classes. Based on this viewpoint, the geometric mean for subspace selection is studied in this paper. Three criteria are analyzed: 1) maximization of the geometric mean of the KL divergences, 2) maximization of the geometric mean of the normalized KL divergences, and 3) the combination of 1 and 2. Preliminary experimental results based on synthetic data, UCI Machine Learning Repository, and handwriting digits show that the third criterion is a potential discriminative subspace selection method, which significantly reduces the class separation problem in comparing with the linear dimensionality reduction step in FLDA and its several representative extensions.
子空间选择方法是模式分类和数据可视化中的强大工具。最重要的子空间方法之一是Fisher线性判别分析(FLDA)中的线性降维步骤,该步骤已成功应用于生物识别、生物信息学和多媒体信息管理等许多领域。然而,FLDA中的线性降维步骤有一个关键缺点:对于具有c个类别的分类任务,如果投影子空间的维度严格低于c - 1,则投影到子空间往往会合并那些在原始特征空间中接近的类。如果从高斯分布中采样不同的类,且所有类具有相同的协方差矩阵,那么FLDA中的线性降维步骤会最大化不同类之间的Kullback-Leibler(KL)散度的平均值。基于此观点,本文研究了用于子空间选择的几何均值。分析了三个标准:1)最大化KL散度的几何均值,2)最大化归一化KL散度的几何均值,以及3)1和2的组合。基于合成数据、UCI机器学习库和手写数字的初步实验结果表明,第三个标准是一种潜在的判别性子空间选择方法,与FLDA中的线性降维步骤及其几个代表性扩展相比,它显著减少了类分离问题。