Suppr超能文献

EPC-DARTS:高效的部分通道连接可微分架构搜索。

EPC-DARTS: Efficient partial channel connection for differentiable architecture search.

机构信息

School of Mathematics and Statistics, Guangdong University of Technology, Guangzhou, China.

School of Mathematics and Statistics, Guangdong University of Technology, Guangzhou, China.

出版信息

Neural Netw. 2023 Sep;166:344-353. doi: 10.1016/j.neunet.2023.07.029. Epub 2023 Jul 24.

Abstract

With weight-sharing and continuous relaxation strategies, the differentiable architecture search (DARTS) proposes a fast and effective solution to perform neural network architecture search in various deep learning tasks. However, unresolved issues, such as the inefficient memory utilization, and the poor stability of the search architecture due to channels randomly selected, which has even caused performance collapses, are still perplexing researchers and practitioners. In this paper, a novel efficient channel attention mechanism based on partial channel connection for differentiable neural architecture search, termed EPC-DARTS, is proposed to address these two issues. Specifically, we design an efficient channel attention module, which is applied to capture cross-channel interactions and assign weight based on channel importance, to dramatically improve search efficiency and reduce memory occupation. Moreover, only partial channels with higher weights in the mixed calculation of operation are used through the efficient channel attention mechanism, and thus unstable network architectures obtained by the random selection operation can also be avoided in the proposed EPC-DARTS. Experimental results show that the proposed EPC-DARTS achieves remarkably competitive performance (CIFAR-10/CIFAR-100: a test accuracy rate of 97.60%/84.02%), compared to other state-of-the-art NAS methods using only 0.2 GPU-Days.

摘要

通过权重共享和连续松弛策略,可微分架构搜索(DARTS)为各种深度学习任务中的神经网络架构搜索提供了一种快速而有效的解决方案。然而,一些未解决的问题仍然困扰着研究人员和从业者,例如效率低下的内存利用问题,以及由于随机选择通道而导致搜索架构不稳定的问题,甚至导致性能崩溃。在本文中,我们提出了一种新颖的基于部分通道连接的可微分神经架构搜索的高效通道注意力机制(EPC-DARTS),以解决这两个问题。具体来说,我们设计了一种高效的通道注意力模块,用于捕获跨通道交互,并根据通道重要性分配权重,从而显著提高搜索效率和减少内存占用。此外,通过高效的通道注意力机制,仅使用操作混合计算中权重较高的部分通道,从而避免了随机选择操作获得的不稳定网络架构。实验结果表明,与其他仅使用 0.2 GPU 天的最先进的 NAS 方法相比,所提出的 EPC-DARTS 实现了卓越的竞争性能(CIFAR-10/CIFAR-100:测试准确率为 97.60%/84.02%)。

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验