Google Research, Mountain View, CA 94043;
Google Research, Mountain View, CA 94043.
Proc Natl Acad Sci U S A. 2019 Jun 11;116(24):11624-11629. doi: 10.1073/pnas.1820657116. Epub 2019 May 24.
Deep neural networks have achieved state-of-the-art accuracy at classifying molecules with respect to whether they bind to specific protein targets. A key breakthrough would occur if these models could reveal the fragment pharmacophores that are causally involved in binding. Extracting chemical details of binding from the networks could enable scientific discoveries about the mechanisms of drug actions. However, doing so requires shining light into the black box that is the trained neural network model, a task that has proved difficult across many domains. Here we show how the binding mechanism learned by deep neural network models can be interrogated, using a recently described attribution method. We first work with carefully constructed synthetic datasets, in which the molecular features responsible for "binding" are fully known. We find that networks that achieve perfect accuracy on held-out test datasets still learn spurious correlations, and we are able to exploit this nonrobustness to construct adversarial examples that fool the model. This makes these models unreliable for accurately revealing information about the mechanisms of protein-ligand binding. In light of our findings, we prescribe a test that checks whether a hypothesized mechanism can be learned. If the test fails, it indicates that the model must be simplified or regularized and/or that the training dataset requires augmentation.
深度神经网络在针对特定蛋白质靶标是否结合的分子分类方面取得了最先进的准确性。如果这些模型能够揭示出与结合因果相关的片段药效团,那么将发生关键突破。从网络中提取结合的化学细节可以使我们对药物作用机制有科学发现。然而,要做到这一点,需要深入研究经过训练的神经网络模型这个“黑箱”,这在许多领域都被证明是困难的。在这里,我们展示了如何使用最近描述的归因方法来询问深度神经网络模型学习的结合机制。我们首先使用精心构建的合成数据集,其中负责“结合”的分子特征是完全已知的。我们发现,在保留的测试数据集上达到完美准确性的网络仍然会学习到虚假相关性,并且我们能够利用这种不稳健性来构建欺骗模型的对抗示例。这使得这些模型无法准确揭示蛋白质 - 配体结合机制的信息。鉴于我们的发现,我们规定了一个检查假设机制是否可以学习的测试。如果测试失败,则表示模型必须简化或正则化,和/或训练数据集需要扩充。