Yang Le, Zheng Ziwei, Han Yizeng, Song Shiji, Huang Gao, Li Fan
IEEE Trans Cybern. 2024 Nov;54(11):6559-6572. doi: 10.1109/TCYB.2024.3455760. Epub 2024 Oct 30.
Differentiable architecture search (DARTS) has emerged as a promising technique for effective neural architecture search, and it mainly contains two steps to find the high-performance architecture. First, the DARTS supernet that consists of mixed operations will be optimized via gradient descent. Second, the final architecture will be built by the selected operations that contribute the most to the supernet. Although DARTS improves the efficiency of neural architecture search (NAS), it suffers from the well-known degeneration issue which can lead to deteriorating architectures. Existing works mainly attribute the degeneration issue to the failure of its supernet optimization, while little attention has been paid to the selection method. In this article, we cease to apply the widely-used magnitude-based selection method and propose a novel criterion based on operation strength that estimates the importance of an operation by its effect on the final loss. We show that the degeneration issue can be effectively addressed by using the proposed criterion without any modification of supernet optimization, indicating that the magnitude-based selection method can be a critical reason for the instability of DARTS. The experiments on NAS-Bench-201 and DARTS search spaces show the effectiveness of our method.
可微架构搜索(DARTS)已成为一种用于有效神经架构搜索的有前途的技术,它主要包含两个步骤来找到高性能架构。首先,由混合操作组成的DARTS超级网络将通过梯度下降进行优化。其次,最终架构将由对超级网络贡献最大的选定操作构建。尽管DARTS提高了神经架构搜索(NAS)的效率,但它存在众所周知的退化问题,这可能导致架构恶化。现有工作主要将退化问题归因于其超级网络优化的失败,而对选择方法关注甚少。在本文中,我们不再应用广泛使用的基于幅度的选择方法,而是提出了一种基于操作强度的新准则,该准则通过操作对最终损失的影响来估计操作的重要性。我们表明,使用所提出的准则可以有效解决退化问题,而无需对超级网络优化进行任何修改,这表明基于幅度的选择方法可能是DARTS不稳定的关键原因。在NAS-Bench-201和DARTS搜索空间上的实验证明了我们方法的有效性。