• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

在 CNN 中推广池化函数:混合、门控和树型。

Generalizing Pooling Functions in CNNs: Mixed, Gated, and Tree.

出版信息

IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):863-875. doi: 10.1109/TPAMI.2017.2703082. Epub 2017 May 12.

DOI:10.1109/TPAMI.2017.2703082
PMID:28504932
Abstract

In this paper, we seek to improve deep neural networks by generalizing the pooling operations that play a central role in the current architectures. We pursue a careful exploration of approaches to allow pooling to learn and to adapt to complex and variable patterns. The two primary directions lie in: (1) learning a pooling function via (two strategies of) combining of max and average pooling, and (2) learning a pooling function in the form of a tree-structured fusion of pooling filters that are themselves learned. In our experiments every generalized pooling operation we explore improves performance when used in place of average or max pooling. We experimentally demonstrate that the proposed pooling operations provide a boost in invariance properties relative to conventional pooling and set the state of the art on several widely adopted benchmark datasets. These benefits come with only a light increase in computational overhead during training (ranging from additional 5 to 15 percent in time complexity) and a very modest increase in the number of model parameters (e.g., additional 1, 9, and 27 parameters for mixed, gated, and 2-level tree pooling operators, respectively). To gain more insights about our proposed pooling methods, we also visualize the learned pooling masks and the embeddings of the internal feature responses for different pooling operations. Our proposed pooling operations are easy to implement and can be applied within various deep neural network architectures.

摘要

在本文中,我们试图通过泛化在当前架构中起核心作用的池化操作来改进深度神经网络。我们仔细探索了允许池化学习和适应复杂和多变模式的方法。两个主要方向在于:(1) 通过(两种组合策略)组合最大池化和平均池化来学习池化函数,(2) 以学习的池化滤波器的树状融合的形式学习池化函数。在我们的实验中,探索的每一种广义池化操作都可以提高性能,替代平均池化或最大池化使用。我们通过实验证明,与传统池化相比,所提出的池化操作提供了对不变性特性的提升,并在几个广泛采用的基准数据集上设定了最新水平。这些好处仅在训练期间增加了轻微的计算开销(时间复杂度增加 5%到 15%),并且模型参数数量仅略有增加(例如,混合、门控和 2 级树状池化操作分别增加 1、9 和 27 个参数)。为了更深入地了解我们提出的池化方法,我们还可视化了学习到的池化掩模和不同池化操作的内部特征响应的嵌入。我们提出的池化操作易于实现,可以应用于各种深度神经网络架构中。

相似文献

1
Generalizing Pooling Functions in CNNs: Mixed, Gated, and Tree.在 CNN 中推广池化函数:混合、门控和树型。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):863-875. doi: 10.1109/TPAMI.2017.2703082. Epub 2017 May 12.
2
A improved pooling method for convolutional neural networks.一种用于卷积神经网络的改进池化方法。
Sci Rep. 2024 Jan 18;14(1):1589. doi: 10.1038/s41598-024-51258-6.
3
Rank-based pooling for deep convolutional neural networks.深度卷积神经网络的基于排序的池化
Neural Netw. 2016 Nov;83:21-31. doi: 10.1016/j.neunet.2016.07.003. Epub 2016 Jul 20.
4
Deep CNNs with Robust LBP Guiding Pooling for Face Recognition.基于稳健局部二值模式引导池化的深度卷积神经网络人脸识别。
Sensors (Basel). 2018 Nov 10;18(11):3876. doi: 10.3390/s18113876.
5
Towards dropout training for convolutional neural networks.面向卷积神经网络的随机失活训练
Neural Netw. 2015 Nov;71:1-10. doi: 10.1016/j.neunet.2015.07.007. Epub 2015 Jul 29.
6
EXAM: A Framework of Learning Extreme and Moderate Embeddings for Person Re-ID.EXAM:一种用于行人重识别的学习极端和适度嵌入的框架。
J Imaging. 2021 Jan 7;7(1):6. doi: 10.3390/jimaging7010006.
7
Hierarchical Recurrent Neural Hashing for Image Retrieval With Hierarchical Convolutional Features.基于层次卷积特征的层次递归神经网络哈希图像检索
IEEE Trans Image Process. 2018;27(1):106-120. doi: 10.1109/TIP.2017.2755766.
8
Convolutional neural network architectures for predicting DNA-protein binding.用于预测DNA-蛋白质结合的卷积神经网络架构。
Bioinformatics. 2016 Jun 15;32(12):i121-i127. doi: 10.1093/bioinformatics/btw255.
9
A sparsity-based stochastic pooling mechanism for deep convolutional neural networks.基于稀疏性的深度卷积神经网络随机池化机制。
Neural Netw. 2018 Sep;105:340-345. doi: 10.1016/j.neunet.2018.05.015. Epub 2018 Jun 15.
10
Deep Convolutional Neural Networks for large-scale speech tasks.用于大规模语音任务的深度卷积神经网络。
Neural Netw. 2015 Apr;64:39-48. doi: 10.1016/j.neunet.2014.08.005. Epub 2014 Sep 16.

引用本文的文献

1
Emotion Recognition from rPPG via Physiologically Inspired Temporal Encoding and Attention-Based Curriculum Learning.基于生理启发的时间编码和基于注意力的课程学习从远程光电容积脉搏波中进行情绪识别。
Sensors (Basel). 2025 Jun 26;25(13):3995. doi: 10.3390/s25133995.
2
Machine Learning Techniques to Infer Protein Structure and Function from Sequences: A Comprehensive Review.基于序列推断蛋白质结构和功能的机器学习技术:全面综述。
Methods Mol Biol. 2025;2867:79-104. doi: 10.1007/978-1-0716-4196-5_5.
3
Recognition of Daily Activities in Adults With Wearable Inertial Sensors: Deep Learning Methods Study.
利用可穿戴惯性传感器识别成年人的日常活动:深度学习方法研究
JMIR Med Inform. 2024 Aug 9;12:e57097. doi: 10.2196/57097.
4
Target Classification Method of Tactile Perception Data with Deep Learning.基于深度学习的触觉感知数据目标分类方法
Entropy (Basel). 2021 Nov 18;23(11):1537. doi: 10.3390/e23111537.
5
Deep Residual Network in Network.网络中的深度残差网络。
Comput Intell Neurosci. 2021 Feb 23;2021:6659083. doi: 10.1155/2021/6659083. eCollection 2021.
6
Recognition of Pashto Handwritten Characters Based on Deep Learning.基于深度学习的普什图文手写字符识别。
Sensors (Basel). 2020 Oct 17;20(20):5884. doi: 10.3390/s20205884.