Zhang Xiao, Liu Xia, Yang Yanyan
Department of Applied Mathematics, School of Sciences, Xi'an University of Technology, Xi'an 710048, China.
Department of Automation, Tsinghua University, Beijing 100084, China.
Entropy (Basel). 2018 Oct 13;20(10):788. doi: 10.3390/e20100788.
The information entropy developed by Shannon is an effective measure of uncertainty in data, and the rough set theory is a useful tool of computer applications to deal with vagueness and uncertainty data circumstances. At present, the information entropy has been extensively applied in the rough set theory, and different information entropy models have also been proposed in rough sets. In this paper, based on the existing feature selection method by using a fuzzy rough set-based information entropy, a corresponding fast algorithm is provided to achieve efficient implementation, in which the fuzzy rough set-based information entropy taking as the evaluation measure for selecting features is computed by an improved mechanism with lower complexity. The essence of the acceleration algorithm is to use iterative reduced instances to compute the lambda-conditional entropy. Numerical experiments are further conducted to show the performance of the proposed fast algorithm, and the results demonstrate that the algorithm acquires the same feature subset to its original counterpart, but with significantly less time.
香农提出的信息熵是衡量数据不确定性的一种有效方法,而粗糙集理论是计算机应用中处理模糊和不确定性数据情况的有用工具。目前,信息熵已在粗糙集理论中得到广泛应用,并且在粗糙集中也提出了不同的信息熵模型。本文基于现有的利用模糊粗糙集信息熵的特征选择方法,提供了一种相应的快速算法以实现高效执行,其中以模糊粗糙集信息熵作为特征选择的评估度量,通过一种复杂度较低的改进机制进行计算。加速算法的本质是使用迭代约简实例来计算λ条件熵。进一步进行了数值实验以展示所提出快速算法的性能,结果表明该算法获得了与原始算法相同的特征子集,但时间显著减少。