Suppr超能文献

部分连接的神经架构搜索,以减少计算冗余。

Partially-Connected Neural Architecture Search for Reduced Computational Redundancy.

出版信息

IEEE Trans Pattern Anal Mach Intell. 2021 Sep;43(9):2953-2970. doi: 10.1109/TPAMI.2021.3059510. Epub 2021 Aug 4.

Abstract

Differentiable architecture search (DARTS) enables effective neural architecture search (NAS) using gradient descent, but suffers from high memory and computational costs. In this paper, we propose a novel approach, namely Partially-Connected DARTS (PC-DARTS), to achieve efficient and stable neural architecture search by reducing the channel and spatial redundancies of the super-network. In the channel level, partial channel connection is presented to randomly sample a small subset of channels for operation selection to accelerate the search process and suppress the over-fitting of the super-network. Side operation is introduced for bypassing (non-sampled) channels to guarantee the performance of searched architectures under extremely low sampling rates. In the spatial level, input features are down-sampled to eliminate spatial redundancy and enhance the efficiency of the mixed computation for operation selection. Furthermore, edge normalization is developed to maintain the consistency of edge selection based on channel sampling with the architectural parameters for edges. Theoretical analysis shows that partial channel connection and parameterized side operation are equivalent to regularizing the super-network on the weights and architectural parameters during bilevel optimization. Experimental results demonstrate that the proposed approach achieves higher search speed and training stability than DARTS. PC-DARTS obtains a top-1 error rate of 2.55 percent on CIFAR-10 with 0.07 GPU-days for architecture search, and a state-of-the-art top-1 error rate of 24.1 percent on ImageNet (under the mobile setting) within 2.8 GPU-days.

摘要

可微分架构搜索(DARTS)利用梯度下降实现了有效的神经架构搜索(NAS),但存在内存和计算成本高的问题。在本文中,我们提出了一种新的方法,即部分连接 DARTS(PC-DARTS),通过减少超级网络的通道和空间冗余来实现高效和稳定的神经架构搜索。在通道级别,提出了部分通道连接,随机选择一小部分通道进行操作选择,以加速搜索过程并抑制超级网络的过拟合。引入旁路操作绕过(未采样)通道,以保证在极低采样率下搜索到的架构的性能。在空间级别,下采样输入特征以消除空间冗余并提高操作选择的混合计算效率。此外,开发了边缘归一化来保持基于通道采样的边缘选择与边缘的架构参数之间的一致性。理论分析表明,部分通道连接和参数化旁路操作在双层优化中相当于对超级网络的权重和架构参数进行正则化。实验结果表明,所提出的方法在搜索速度和训练稳定性方面优于 DARTS。PC-DARTS 在 CIFAR-10 上实现了 2.55%的 top-1 错误率,搜索架构的 GPU 时间为 0.07 GPU 天,在 ImageNet 上(在移动设置下)实现了 24.1%的最先进的 top-1 错误率,GPU 时间为 2.8 GPU 天。

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验