Suppr超能文献

OStr-DARTS:基于操作强度的可微神经架构搜索

OStr-DARTS: Differentiable Neural Architecture Search Based on Operation Strength.

作者信息

Yang Le, Zheng Ziwei, Han Yizeng, Song Shiji, Huang Gao, Li Fan

出版信息

IEEE Trans Cybern. 2024 Nov;54(11):6559-6572. doi: 10.1109/TCYB.2024.3455760. Epub 2024 Oct 30.

Abstract

Differentiable architecture search (DARTS) has emerged as a promising technique for effective neural architecture search, and it mainly contains two steps to find the high-performance architecture. First, the DARTS supernet that consists of mixed operations will be optimized via gradient descent. Second, the final architecture will be built by the selected operations that contribute the most to the supernet. Although DARTS improves the efficiency of neural architecture search (NAS), it suffers from the well-known degeneration issue which can lead to deteriorating architectures. Existing works mainly attribute the degeneration issue to the failure of its supernet optimization, while little attention has been paid to the selection method. In this article, we cease to apply the widely-used magnitude-based selection method and propose a novel criterion based on operation strength that estimates the importance of an operation by its effect on the final loss. We show that the degeneration issue can be effectively addressed by using the proposed criterion without any modification of supernet optimization, indicating that the magnitude-based selection method can be a critical reason for the instability of DARTS. The experiments on NAS-Bench-201 and DARTS search spaces show the effectiveness of our method.

摘要

可微架构搜索(DARTS)已成为一种用于有效神经架构搜索的有前途的技术,它主要包含两个步骤来找到高性能架构。首先,由混合操作组成的DARTS超级网络将通过梯度下降进行优化。其次,最终架构将由对超级网络贡献最大的选定操作构建。尽管DARTS提高了神经架构搜索(NAS)的效率,但它存在众所周知的退化问题,这可能导致架构恶化。现有工作主要将退化问题归因于其超级网络优化的失败,而对选择方法关注甚少。在本文中,我们不再应用广泛使用的基于幅度的选择方法,而是提出了一种基于操作强度的新准则,该准则通过操作对最终损失的影响来估计操作的重要性。我们表明,使用所提出的准则可以有效解决退化问题,而无需对超级网络优化进行任何修改,这表明基于幅度的选择方法可能是DARTS不稳定的关键原因。在NAS-Bench-201和DARTS搜索空间上的实验证明了我们方法的有效性。

相似文献

1
OStr-DARTS: Differentiable Neural Architecture Search Based on Operation Strength.OStr-DARTS:基于操作强度的可微神经架构搜索
IEEE Trans Cybern. 2024 Nov;54(11):6559-6572. doi: 10.1109/TCYB.2024.3455760. Epub 2024 Oct 30.
2
Improving Differentiable Architecture Search via self-distillation.通过自蒸馏改进可微架构搜索。
Neural Netw. 2023 Oct;167:656-667. doi: 10.1016/j.neunet.2023.08.062. Epub 2023 Sep 9.
4
Partially-Connected Neural Architecture Search for Reduced Computational Redundancy.部分连接的神经架构搜索,以减少计算冗余。
IEEE Trans Pattern Anal Mach Intell. 2021 Sep;43(9):2953-2970. doi: 10.1109/TPAMI.2021.3059510. Epub 2021 Aug 4.
8
A Gradient-Guided Evolutionary Neural Architecture Search.一种梯度引导的进化神经网络架构搜索。
IEEE Trans Neural Netw Learn Syst. 2025 Mar;36(3):4345-4357. doi: 10.1109/TNNLS.2024.3371432. Epub 2025 Feb 28.
9
Neural Architecture Selection as a Nash Equilibrium With Batch Entanglement.作为具有批量纠缠的纳什均衡的神经架构选择
IEEE Trans Neural Netw Learn Syst. 2024 Nov;35(11):15195-15209. doi: 10.1109/TNNLS.2023.3283239. Epub 2024 Oct 29.
10
Cyclic Differentiable Architecture Search.循环可微架构搜索
IEEE Trans Pattern Anal Mach Intell. 2023 Jan;45(1):211-228. doi: 10.1109/TPAMI.2022.3153065. Epub 2022 Dec 5.

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验