• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

为不同层自适应定制激活函数

Adaptively Customizing Activation Functions for Various Layers.

作者信息

Hu Haigen, Liu Aizhu, Guan Qiu, Qian Hanwang, Li Xiaoxin, Chen Shengyong, Zhou Qianwei

出版信息

IEEE Trans Neural Netw Learn Syst. 2023 Sep;34(9):6096-6107. doi: 10.1109/TNNLS.2021.3133263. Epub 2023 Sep 1.

DOI:10.1109/TNNLS.2021.3133263
PMID:35007200
Abstract

To enhance the nonlinearity of neural networks and increase their mapping abilities between the inputs and response variables, activation functions play a crucial role to model more complex relationships and patterns in the data. In this work, a novel methodology is proposed to adaptively customize activation functions only by adding very few parameters to the traditional activation functions such as Sigmoid, Tanh, and rectified linear unit (ReLU). To verify the effectiveness of the proposed methodology, some theoretical and experimental analysis on accelerating the convergence and improving the performance is presented, and a series of experiments are conducted based on various network models (such as AlexNet, VggNet, GoogLeNet, ResNet and DenseNet), and various datasets (such as CIFAR10, CIFAR100, miniImageNet, PASCAL VOC, and COCO). To further verify the validity and suitability in various optimization strategies and usage scenarios, some comparison experiments are also implemented among different optimization strategies (such as SGD, Momentum, AdaGrad, AdaDelta, and ADAM) and different recognition tasks such as classification and detection. The results show that the proposed methodology is very simple but with significant performance in convergence speed, precision, and generalization, and it can surpass other popular methods such as ReLU and adaptive functions such as Swish in almost all experiments in terms of overall performance.

摘要

为了增强神经网络的非线性并提高其在输入和响应变量之间的映射能力,激活函数对于在数据中建模更复杂的关系和模式起着至关重要的作用。在这项工作中,提出了一种新颖的方法,仅通过在诸如Sigmoid、Tanh和整流线性单元(ReLU)等传统激活函数上添加极少的参数来自适应地定制激活函数。为了验证所提方法的有效性,给出了一些关于加速收敛和提高性能的理论和实验分析,并基于各种网络模型(如AlexNet、VggNet、GoogLeNet、ResNet和DenseNet)以及各种数据集(如CIFAR10、CIFAR100、miniImageNet、PASCAL VOC和COCO)进行了一系列实验。为了进一步验证在各种优化策略和使用场景中的有效性和适用性,还在不同的优化策略(如SGD、Momentum、AdaGrad、AdaDelta和ADAM)以及不同的识别任务(如分类和检测)之间进行了一些比较实验。结果表明,所提方法非常简单,但在收敛速度、精度和泛化方面具有显著性能,并且在几乎所有实验的整体性能方面都能超越其他流行方法(如ReLU)和自适应函数(如Swish)。

相似文献

1
Adaptively Customizing Activation Functions for Various Layers.为不同层自适应定制激活函数
IEEE Trans Neural Netw Learn Syst. 2023 Sep;34(9):6096-6107. doi: 10.1109/TNNLS.2021.3133263. Epub 2023 Sep 1.
2
diffGrad: An Optimization Method for Convolutional Neural Networks.差分梯度法:卷积神经网络的一种优化方法。
IEEE Trans Neural Netw Learn Syst. 2020 Nov;31(11):4500-4511. doi: 10.1109/TNNLS.2019.2955777. Epub 2020 Oct 29.
3
Parametric Deformable Exponential Linear Units for deep neural networks.参数变形指数线性单元在深度神经网络中的应用。
Neural Netw. 2020 May;125:281-289. doi: 10.1016/j.neunet.2020.02.012. Epub 2020 Feb 26.
4
Re-Thinking the Effectiveness of Batch Normalization and Beyond.重新思考批归一化的有效性及其他
IEEE Trans Pattern Anal Mach Intell. 2024 Jan;46(1):465-478. doi: 10.1109/TPAMI.2023.3319005. Epub 2023 Dec 5.
5
The WuC-Adam algorithm based on joint improvement of Warmup and cosine annealing algorithms.基于热身算法和余弦退火算法联合改进的WuC-Adam算法。
Math Biosci Eng. 2024 Jan;21(1):1270-1285. doi: 10.3934/mbe.2024054. Epub 2022 Dec 26.
6
A novel scaled-gamma-tanh (SGT) activation function in 3D CNN applied for MRI classification.一种应用于 MRI 分类的三维卷积神经网络中新的尺度伽马双曲正切(SGT)激活函数。
Sci Rep. 2022 Sep 2;12(1):14978. doi: 10.1038/s41598-022-19020-y.
7
Optimization of Microchannels and Application of Basic Activation Functions of Deep Neural Network for Accuracy Analysis of Microfluidic Parameter Data.用于微流体参数数据准确性分析的微通道优化及深度神经网络基本激活函数的应用
Micromachines (Basel). 2022 Aug 20;13(8):1352. doi: 10.3390/mi13081352.
8
Discovering Parametric Activation Functions.发现参数激活函数。
Neural Netw. 2022 Apr;148:48-65. doi: 10.1016/j.neunet.2022.01.001. Epub 2022 Jan 7.
9
A novel adaptive momentum method for medical image classification using convolutional neural network.基于卷积神经网络的医学图像分类自适应动量方法
BMC Med Imaging. 2022 Mar 1;22(1):34. doi: 10.1186/s12880-022-00755-z.
10
Automated detection of leukemia by pretrained deep neural networks and transfer learning: A comparison.基于预训练深度神经网络和迁移学习的白血病自动检测:比较研究。
Med Eng Phys. 2021 Dec;98:8-19. doi: 10.1016/j.medengphy.2021.10.006. Epub 2021 Oct 13.