• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

结合多分支卷积神经网络和改进型变压器的进化神经架构搜索

Evolutionary neural architecture search combining multi-branch ConvNet and improved transformer.

作者信息

Xu Yang, Ma Yongjie

机构信息

College of Physics and Electronic Engineering, Northwest Normal University, Lanzhou, 730070, China.

出版信息

Sci Rep. 2023 Sep 22;13(1):15791. doi: 10.1038/s41598-023-42931-3.

DOI:10.1038/s41598-023-42931-3
PMID:37737271
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10516961/
Abstract

Deep convolutional neural networks (CNNs) have achieved promising performance in the field of deep learning, but the manual design turns out to be very difficult due to the increasingly complex topologies of CNNs. Recently, neural architecture search (NAS) methods have been proposed to automatically design network architectures, which are superior to handcrafted counterparts. Unfortunately, most current NAS methods suffer from either highly computational complexity of generated architectures or limitations in the flexibility of architecture design. To address above issues, this article proposes an evolutionary neural architecture search (ENAS) method based on improved Transformer and multi-branch ConvNet. The multi-branch block enriches the feature space and enhances the representational capacity of a network by combining paths with different complexities. Since convolution is inherently a local operation, a simple yet powerful "batch-free normalization Transformer Block" (BFNTBlock) is proposed to leverage both local information and long-range feature dependencies. In particular, the design of batch-free normalization (BFN) and batch normalization (BN) mixed in the BFNTBlock blocks the accumulation of estimation shift ascribe to the stack of BN, which has favorable effects for performance improvement. The proposed method achieves remarkable accuracies, 97.24 [Formula: see text] and 80.06 [Formula: see text] on CIFAR10 and CIFAR100, respectively, with high computational efficiency, i.e. only 1.46 and 1.53 GPU days. To validate the universality of our method in application scenarios, the proposed algorithm is verified on two real-world applications, including the GTSRB and NEU-CLS dataset, and achieves a better performance than common methods.

摘要

深度卷积神经网络(CNN)在深度学习领域取得了令人瞩目的成绩,但由于CNN的拓扑结构日益复杂,手动设计变得非常困难。最近,神经架构搜索(NAS)方法被提出来自动设计网络架构,这些方法优于手工设计的架构。不幸的是,当前大多数NAS方法要么存在生成架构的计算复杂度高的问题,要么在架构设计的灵活性方面存在局限性。为了解决上述问题,本文提出了一种基于改进的Transformer和多分支卷积网络的进化神经架构搜索(ENAS)方法。多分支模块通过组合具有不同复杂度的路径来丰富特征空间并增强网络的表示能力。由于卷积本质上是一种局部操作,因此提出了一种简单而强大的“无批归一化Transformer模块”(BFNTBlock),以利用局部信息和长距离特征依赖关系。特别是,BFNTBlock中无批归一化(BFN)和批归一化(BN)混合的设计阻止了由于BN堆叠导致的估计偏移的累积,这对性能提升有积极影响。所提出的方法在CIFAR10和CIFAR100上分别取得了显著的准确率,即97.24 [公式:见原文] 和80.06 [公式:见原文],且计算效率高,即仅需1.46和1.53个GPU日。为了验证我们方法在应用场景中的通用性,该算法在两个实际应用中进行了验证,包括GTSRB和NEU-CLS数据集,并取得了比常用方法更好的性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/bb26ee020b47/41598_2023_42931_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/276126c754e6/41598_2023_42931_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/815568d5af40/41598_2023_42931_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/edfd06aefd3e/41598_2023_42931_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/1fb7de32c195/41598_2023_42931_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/4c6520901d21/41598_2023_42931_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/6074026a719e/41598_2023_42931_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/91576fbd2e90/41598_2023_42931_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/eddc7f9337ac/41598_2023_42931_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/bb26ee020b47/41598_2023_42931_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/276126c754e6/41598_2023_42931_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/815568d5af40/41598_2023_42931_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/edfd06aefd3e/41598_2023_42931_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/1fb7de32c195/41598_2023_42931_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/4c6520901d21/41598_2023_42931_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/6074026a719e/41598_2023_42931_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/91576fbd2e90/41598_2023_42931_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/eddc7f9337ac/41598_2023_42931_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f656/10516961/bb26ee020b47/41598_2023_42931_Fig9_HTML.jpg

相似文献

1
Evolutionary neural architecture search combining multi-branch ConvNet and improved transformer.结合多分支卷积神经网络和改进型变压器的进化神经架构搜索
Sci Rep. 2023 Sep 22;13(1):15791. doi: 10.1038/s41598-023-42931-3.
2
Surrogate-Assisted Particle Swarm Optimization for Evolving Variable-Length Transferable Blocks for Image Classification.用于图像分类的可变长度可转移块进化的代理辅助粒子群优化算法
IEEE Trans Neural Netw Learn Syst. 2022 Aug;33(8):3727-3740. doi: 10.1109/TNNLS.2021.3054400. Epub 2022 Aug 3.
3
MBUTransNet: multi-branch U-shaped network fusion transformer architecture for medical image segmentation.MBUTransNet:用于医学图像分割的多分支 U 形网络融合变压器架构。
Int J Comput Assist Radiol Surg. 2023 Oct;18(10):1895-1902. doi: 10.1007/s11548-023-02879-1. Epub 2023 Apr 6.
4
Enhanced Gradient for Differentiable Architecture Search.用于可微架构搜索的增强梯度
IEEE Trans Neural Netw Learn Syst. 2024 Jul;35(7):9606-9620. doi: 10.1109/TNNLS.2023.3235479. Epub 2024 Jul 8.
5
Automated multi-modal Transformer network (AMTNet) for 3D medical images segmentation.用于3D医学图像分割的自动多模态Transformer网络(AMTNet)。
Phys Med Biol. 2023 Jan 9;68(2). doi: 10.1088/1361-6560/aca74c.
6
A Gradient-Guided Evolutionary Neural Architecture Search.一种梯度引导的进化神经网络架构搜索。
IEEE Trans Neural Netw Learn Syst. 2025 Mar;36(3):4345-4357. doi: 10.1109/TNNLS.2024.3371432. Epub 2025 Feb 28.
7
A Cell-Based Fast Memetic Algorithm for Automated Convolutional Neural Architecture Design.一种基于细胞的快速 Memetic 算法用于自动卷积神经网络架构设计。
IEEE Trans Neural Netw Learn Syst. 2023 Nov;34(11):9040-9053. doi: 10.1109/TNNLS.2022.3155230. Epub 2023 Oct 27.
8
AUTO-HAR: An adaptive human activity recognition framework using an automated CNN architecture design.AUTO-HAR:一种使用自动化卷积神经网络(CNN)架构设计的自适应人类活动识别框架。
Heliyon. 2023 Feb 13;9(2):e13636. doi: 10.1016/j.heliyon.2023.e13636. eCollection 2023 Feb.
9
A Fast Evolutionary Knowledge Transfer Search for Multiscale Deep Neural Architecture.一种用于多尺度深度神经架构的快速进化知识转移搜索
IEEE Trans Neural Netw Learn Syst. 2024 Dec;35(12):17450-17464. doi: 10.1109/TNNLS.2023.3304291. Epub 2024 Dec 2.
10
ESC-NAS: Environment Sound Classification Using Hardware-Aware Neural Architecture Search for the Edge.ESC-NAS:利用硬件感知神经架构搜索进行边缘环境声音分类
Sensors (Basel). 2024 Jun 9;24(12):3749. doi: 10.3390/s24123749.

引用本文的文献

1
Nature-Inspired Intelligent Computing: A Comprehensive Survey.受自然启发的智能计算:全面综述
Research (Wash D C). 2024 Aug 16;7:0442. doi: 10.34133/research.0442. eCollection 2024.
2
Optimizing Image Classification: Automated Deep Learning Architecture Crafting with Network and Learning Hyperparameter Tuning.优化图像分类:通过网络和学习超参数调整实现自动深度学习架构构建
Biomimetics (Basel). 2023 Nov 4;8(7):525. doi: 10.3390/biomimetics8070525.

本文引用的文献

1
Normalization Techniques in Training DNNs: Methodology, Analysis and Application.深度学习神经网络训练中的归一化技术:方法、分析与应用。
IEEE Trans Pattern Anal Mach Intell. 2023 Aug;45(8):10173-10196. doi: 10.1109/TPAMI.2023.3250241. Epub 2023 Jun 30.
2
A Survey on Evolutionary Neural Architecture Search.进化神经架构搜索综述
IEEE Trans Neural Netw Learn Syst. 2023 Feb;34(2):550-570. doi: 10.1109/TNNLS.2021.3100554. Epub 2023 Feb 3.
3
Surrogate-Assisted Particle Swarm Optimization for Evolving Variable-Length Transferable Blocks for Image Classification.
用于图像分类的可变长度可转移块进化的代理辅助粒子群优化算法
IEEE Trans Neural Netw Learn Syst. 2022 Aug;33(8):3727-3740. doi: 10.1109/TNNLS.2021.3054400. Epub 2022 Aug 3.
4
Automatically Designing CNN Architectures Using the Genetic Algorithm for Image Classification.使用遗传算法自动设计用于图像分类的 CNN 架构。
IEEE Trans Cybern. 2020 Sep;50(9):3840-3854. doi: 10.1109/TCYB.2020.2983860. Epub 2020 Apr 21.
5
BlockQNN: Efficient Block-Wise Neural Network Architecture Generation.BlockQNN:高效的分块神经网络架构生成。
IEEE Trans Pattern Anal Mach Intell. 2021 Jul;43(7):2314-2328. doi: 10.1109/TPAMI.2020.2969193. Epub 2021 Jun 8.
6
Completely Automated CNN Architecture Design Based on Blocks.基于块的完全自动化 CNN 架构设计。
IEEE Trans Neural Netw Learn Syst. 2020 Apr;31(4):1242-1254. doi: 10.1109/TNNLS.2019.2919608. Epub 2019 Jun 20.
7
A Generic Deep-Learning-Based Approach for Automated Surface Inspection.基于深度学习的通用自动表面检测方法。
IEEE Trans Cybern. 2018 Mar;48(3):929-940. doi: 10.1109/TCYB.2017.2668395. Epub 2017 Feb 24.
8
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks.更快的 R-CNN:基于区域建议网络的实时目标检测。
IEEE Trans Pattern Anal Mach Intell. 2017 Jun;39(6):1137-1149. doi: 10.1109/TPAMI.2016.2577031. Epub 2016 Jun 6.
9
Deep learning.深度学习。
Nature. 2015 May 28;521(7553):436-44. doi: 10.1038/nature14539.
10
Multi-column deep neural network for traffic sign classification.多列深度神经网络用于交通标志分类。
Neural Netw. 2012 Aug;32:333-8. doi: 10.1016/j.neunet.2012.02.023. Epub 2012 Feb 14.