Bontempi Gianluca
Départment d'Informative,Université Libre de Bruxelles, Bruxelles, Belgium.
IEEE/ACM Trans Comput Biol Bioinform. 2007 Apr-Jun;4(2):293-300. doi: 10.1109/TCBB.2007.1014.
Because of high dimensionality, machine learning algorithms typically rely on feature selection techniques in order to perform effective classification in microarray gene expression data sets. However, the large number of features compared to the number of samples makes the task of feature selection computationally hard and prone to errors. This paper interprets feature selection as a task of stochastic optimization, where the goal is to select among an exponential number of alternative gene subsets the one expected to return the highest generalization in classification. Blocking is an experimental design strategy which produces similar experimental conditions to compare alternative stochastic configurations in order to be confident that observed differences in accuracy are due to actual differences rather than to fluctuations and noise effects. We propose an original blocking strategy for improving feature selection which aggregates in a paired way the validation outcomes of several learning algorithms to assess a gene subset and compare it to others. This is a novelty with respect to conventional wrappers, which commonly adopt a sole learning algorithm to evaluate the relevance of a given set of variables. The rationale of the approach is that, by increasing the amount of experimental conditions under which we validate a feature subset, we can lessen the problems related to the scarcity of samples and consequently come up with a better selection. The paper shows that the blocking strategy significantly improves the performance of a conventional forward selection for a set of 16 publicly available cancer expression data sets. The experiments involve six different classifiers and show that improvements take place independent of the classification algorithm used after the selection step. Two further validations based on available biological annotation support the claim that blocking strategies in feature selection may improve the accuracy and the quality of the solution. The first validation is based on retrieving PubMEd abstracts associated to the selected genes and matching them to regular expressions describing the biological phenomenon underlying the expression data sets. The biological validation that follows is based on the use of the Bioconductor package GoStats in order to perform Gene Ontology statistical analysis.
由于维度较高,机器学习算法通常依赖特征选择技术,以便在微阵列基因表达数据集中进行有效的分类。然而,与样本数量相比,特征数量众多使得特征选择任务在计算上变得困难且容易出错。本文将特征选择解释为一个随机优化任务,其目标是在指数数量的替代基因子集中选择一个预期在分类中具有最高泛化能力的子集。分块是一种实验设计策略,它产生相似的实验条件以比较替代随机配置,从而确信观察到的准确性差异是由于实际差异而非波动和噪声效应。我们提出一种用于改进特征选择的原创分块策略,该策略以配对方式聚合几种学习算法的验证结果,以评估一个基因子集并将其与其他子集进行比较。这相对于传统包装器来说是一个新颖之处,传统包装器通常采用单一学习算法来评估给定变量集的相关性。该方法的基本原理是,通过增加验证特征子集的实验条件数量,我们可以减少与样本稀缺相关的问题,从而得出更好的选择。本文表明,对于一组16个公开可用的癌症表达数据集,分块策略显著提高了传统前向选择的性能。实验涉及六种不同的分类器,结果表明,改进的发生与选择步骤后使用的分类算法无关。基于可用生物学注释的另外两项验证支持了特征选择中的分块策略可能提高解决方案的准确性和质量这一说法。第一次验证基于检索与所选基因相关的PubMEd摘要,并将它们与描述表达数据集背后生物学现象的正则表达式进行匹配。随后的生物学验证基于使用Bioconductor包GoStats来进行基因本体统计分析。