Koo Peter K, Ploenzke Matt
Simons Center for Quantitative Biology, Cold Spring Harbor Laboratory, Cold Spring Harbor, NY, USA.
Department of Biostatistics, T.H. Chan School of Public Health, Harvard University, Boston, MA, USA.
Nat Mach Intell. 2021 Mar;3(3):258-266. doi: 10.1038/s42256-020-00291-x. Epub 2021 Feb 8.
Deep convolutional neural networks (CNNs) trained on regulatory genomic sequences tend to build representations in a distributed manner, making it a challenge to extract learned features that are biologically meaningful, such as sequence motifs. Here we perform a comprehensive analysis on synthetic sequences to investigate the role that CNN activations have on model interpretability. We show that employing an exponential activation to first layer filters consistently leads to interpretable and robust representations of motifs compared to other commonly used activations. Strikingly, we demonstrate that CNNs with better test performance do not necessarily imply more interpretable representations with attribution methods. We find that CNNs with exponential activations significantly improve the efficacy of recovering biologically meaningful representations with attribution methods. We demonstrate these results generalise to real DNA sequences across several datasets. Together, this work demonstrates how a small modification to existing CNNs, i.e. setting exponential activations in the first layer, can significantly improve the robustness and interpretabilty of learned representations directly in convolutional filters and indirectly with attribution methods.
在调控基因组序列上训练的深度卷积神经网络(CNN)倾向于以分布式方式构建表示,这使得提取具有生物学意义的学习特征(如序列基序)成为一项挑战。在这里,我们对合成序列进行了全面分析,以研究CNN激活在模型可解释性方面所起的作用。我们表明,与其他常用激活函数相比,对第一层滤波器采用指数激活函数始终会产生可解释且稳健的基序表示。引人注目的是,我们证明了具有更好测试性能的CNN并不一定意味着使用归因方法时具有更可解释的表示。我们发现,具有指数激活函数的CNN通过归因方法显著提高了恢复具有生物学意义表示的效率。我们证明这些结果可以推广到多个数据集中的真实DNA序列。总之,这项工作展示了对现有CNN进行的一个小修改,即在第一层设置指数激活函数,如何能够直接在卷积滤波器中以及间接通过归因方法显著提高学习表示的稳健性和可解释性。