Suppr超能文献

Kafnets:基于核的神经网络非参数激活函数。

Kafnets: Kernel-based non-parametric activation functions for neural networks.

机构信息

Department of Information Engineering, Electronics and Telecommunications (DIET), Sapienza University of Rome, Via Eudossiana 18, 00184 Rome, Italy.

Department of Communications Engineering, University of Cantabria, Av. los Castros s/n, 39005 Santander, Cantabria, Spain.

出版信息

Neural Netw. 2019 Feb;110:19-32. doi: 10.1016/j.neunet.2018.11.002. Epub 2018 Nov 13.

Abstract

Neural networks are generally built by interleaving (adaptable) linear layers with (fixed) nonlinear activation functions. To increase their flexibility, several authors have proposed methods for adapting the activation functions themselves, endowing them with varying degrees of flexibility. None of these approaches, however, have gained wide acceptance in practice, and research in this topic remains open. In this paper, we introduce a novel family of flexible activation functions that are based on an inexpensive kernel expansion at every neuron. Leveraging several properties of kernel-based models, we propose multiple variations for designing and initializing these kernel activation functions (KAFs), including a multidimensional scheme allowing to nonlinearly combine information from different paths in the network. The resulting KAFs can approximate any mapping defined over a subset of the real line, either convex or non-convex. Furthermore, they are smooth over their entire domain, linear in their parameters, and they can be regularized using any known scheme, including the use of ℓ penalties to enforce sparseness. To the best of our knowledge, no other known model satisfies all these properties simultaneously. In addition, we provide an overview on alternative techniques for adapting the activation functions, which is currently lacking in the literature. A large set of experiments validates our proposal.

摘要

神经网络通常通过交错(可适应)线性层和(固定)非线性激活函数来构建。为了提高它们的灵活性,一些作者提出了自适应激活函数本身的方法,使它们具有不同程度的灵活性。然而,这些方法都没有在实践中得到广泛的认可,这个领域的研究仍然是开放的。在本文中,我们引入了一种新的灵活激活函数家族,它基于每个神经元的廉价核扩展。利用核模型的几个特性,我们提出了多种设计和初始化这些核激活函数(KAF)的变体,包括一种多维方案,允许非线性地组合网络中不同路径的信息。由此产生的 KAF 可以逼近实数子集上定义的任何映射,无论是凸的还是非凸的。此外,它们在整个域上是平滑的,参数是线性的,并且可以使用任何已知的方案进行正则化,包括使用ℓ惩罚来强制稀疏性。据我们所知,没有其他已知的模型同时满足所有这些特性。此外,我们还提供了一个关于自适应激活函数的替代技术的概述,这在文献中是缺乏的。大量的实验验证了我们的建议。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验