Neural Comput. 2010 Nov;22(11):2887-923. doi: 10.1162/NECO_a_00027.
Reducing the dimensionality of high-dimensional data without losing its essential information is an important task in information processing. When class labels of training data are available, Fisher discriminant analysis (FDA) has been widely used. However, the optimality of FDA is guaranteed only in a very restricted ideal circumstance, and it is often observed that FDA does not provide a good classification surface for many real problems. This letter treats the problem of supervised dimensionality reduction from the viewpoint of information theory and proposes a framework of dimensionality reduction based on class-conditional entropy minimization. The proposed linear dimensionality-reduction technique is validated both theoretically and experimentally. Then, through kernel Fisher discriminant analysis (KFDA), the multiple kernel learning problem is treated in the proposed framework, and a novel algorithm, which iteratively optimizes the parameters of the classification function and kernel combination coefficients, is proposed. The algorithm is experimentally shown to be comparable to or outperforms KFDA for large-scale benchmark data sets, and comparable to other multiple kernel learning techniques on the yeast protein function annotation task.
在信息处理中,降低高维数据的维度而不丢失其基本信息是一项重要任务。当训练数据的类别标签可用时,Fisher 判别分析 (FDA) 已被广泛应用。然而,FDA 的最优性仅在非常受限的理想情况下得到保证,并且经常观察到 FDA 并没有为许多实际问题提供一个良好的分类曲面。这封信从信息论的角度处理监督降维问题,并提出了一种基于类别条件熵最小化的降维框架。所提出的线性降维技术在理论和实验上都得到了验证。然后,通过核 Fisher 判别分析 (KFDA),在提出的框架中处理多核学习问题,并提出了一种新的算法,该算法迭代优化分类函数和核组合系数的参数。实验表明,该算法在大规模基准数据集上与 KFDA 相当或优于 KFDA,在酵母蛋白质功能注释任务上与其他多核学习技术相当。