Koutsoukas Alexios, Monaghan Keith J, Li Xiaoli, Huan Jun
Department of Electrical Engineering and Computer Sciences, University of Kansas, Lawrence, KS, 66047-7621, USA.
J Cheminform. 2017 Jun 28;9(1):42. doi: 10.1186/s13321-017-0226-y.
In recent years, research in artificial neural networks has resurged, now under the deep-learning umbrella, and grown extremely popular. Recently reported success of DL techniques in crowd-sourced QSAR and predictive toxicology competitions has showcased these methods as powerful tools in drug-discovery and toxicology research. The aim of this work was dual, first large number of hyper-parameter configurations were explored to investigate how they affect the performance of DNNs and could act as starting points when tuning DNNs and second their performance was compared to popular methods widely employed in the field of cheminformatics namely Naïve Bayes, k-nearest neighbor, random forest and support vector machines. Moreover, robustness of machine learning methods to different levels of artificially introduced noise was assessed. The open-source Caffe deep-learning framework and modern NVidia GPU units were utilized to carry out this study, allowing large number of DNN configurations to be explored.
We show that feed-forward deep neural networks are capable of achieving strong classification performance and outperform shallow methods across diverse activity classes when optimized. Hyper-parameters that were found to play critical role are the activation function, dropout regularization, number hidden layers and number of neurons. When compared to the rest methods, tuned DNNs were found to statistically outperform, with p value <0.01 based on Wilcoxon statistical test. DNN achieved on average MCC units of 0.149 higher than NB, 0.092 than kNN, 0.052 than SVM with linear kernel, 0.021 than RF and finally 0.009 higher than SVM with radial basis function kernel. When exploring robustness to noise, non-linear methods were found to perform well when dealing with low levels of noise, lower than or equal to 20%, however when dealing with higher levels of noise, higher than 30%, the Naïve Bayes method was found to perform well and even outperform at the highest level of noise 50% more sophisticated methods across several datasets.
近年来,人工神经网络的研究再度兴起,如今在深度学习的范畴下,变得极为流行。近期报道的深度学习技术在众包定量构效关系(QSAR)和预测毒理学竞赛中的成功,已将这些方法展示为药物发现和毒理学研究中的强大工具。这项工作的目标有两个,首先探索大量超参数配置,以研究它们如何影响深度神经网络(DNN)的性能,并在调整DNN时作为起点,其次将其性能与化学信息学领域广泛使用的常用方法(即朴素贝叶斯、k近邻、随机森林和支持向量机)进行比较。此外,还评估了机器学习方法对不同水平人工引入噪声的鲁棒性。利用开源的Caffe深度学习框架和现代英伟达GPU单元来开展这项研究,从而能够探索大量的DNN配置。
我们表明,前馈深度神经网络在经过优化后,能够实现强大的分类性能,并且在各种活性类别上优于浅层方法。被发现起关键作用的超参数是激活函数、随机失活正则化、隐藏层数和神经元数量。与其他方法相比,经调整的DNN在统计学上表现更优,基于威尔科克森统计检验,p值<0.01。DNN平均马修斯相关系数(MCC)单位比朴素贝叶斯高0.149,比k近邻高0.092,比线性核支持向量机高0.052,比随机森林高0.021,最终比径向基函数核支持向量机高0.009。在探索对噪声的鲁棒性时,发现非线性方法在处理低于或等于20%的低水平噪声时表现良好,然而在处理高于30%的高水平噪声时,朴素贝叶斯方法表现良好,甚至在最高噪声水平50%时优于几个数据集中更复杂的方法。