• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

你只需搜索一次:通过直接稀疏优化的单镜头神经架构搜索。

You Only Search Once: Single Shot Neural Architecture Search via Direct Sparse Optimization.

出版信息

IEEE Trans Pattern Anal Mach Intell. 2021 Sep;43(9):2891-2904. doi: 10.1109/TPAMI.2020.3020300. Epub 2021 Aug 4.

DOI:10.1109/TPAMI.2020.3020300
PMID:32866093
Abstract

Recently neural architecture search (NAS) has raised great interest in both academia and industry. However, it remains challenging because of its huge and non-continuous search space. Instead of applying evolutionary algorithm or reinforcement learning as previous works, this paper proposes a direct sparse optimization NAS (DSO-NAS) method. The motivation behind DSO-NAS is to address the task in the view of model pruning. To achieve this goal, we start from a completely connected block, and then introduce scaling factors to scale the information flow between operations. Next, sparse regularizations are imposed to prune useless connections in the architecture. Lastly, an efficient and theoretically sound optimization method is derived to solve it. Our method enjoys both advantages of differentiability and efficiency, therefore it can be directly applied to large datasets like ImageNet and tasks beyond classification. Particularly, on the CIFAR-10 dataset, DSO-NAS achieves an average test error 2.74 percent, while on the ImageNet dataset DSO-NAS achieves 25.4 percent test error under 600M FLOPs with 8 GPUs in 18 hours. As for semantic segmentation task, DSO-NAS also achieve competitive result compared with manually designed architectures on the PASCAL VOC dataset. Code is available at https://github.com/XinbangZhang/DSO-NAS.

摘要

最近,神经架构搜索(NAS)在学术界和工业界都引起了极大的兴趣。然而,由于其巨大的、不连续的搜索空间,它仍然具有挑战性。本文提出了一种直接稀疏优化 NAS(DSO-NAS)方法,而不是像以前的工作那样应用进化算法或强化学习。DSO-NAS 的动机是从模型剪枝的角度来解决这个任务。为了实现这一目标,我们从一个完全连接的块开始,然后引入缩放因子来缩放操作之间的信息流。接下来,施加稀疏正则化来剪枝架构中的无用连接。最后,导出了一种高效且理论上合理的优化方法来解决这个问题。我们的方法既具有可区分性又具有效率,因此可以直接应用于像 ImageNet 这样的大型数据集和分类任务之外的任务。特别是,在 CIFAR-10 数据集上,DSO-NAS 的平均测试误差为 2.74%,而在 ImageNet 数据集上,在 8 个 GPU 上用 6 亿 FLOPs 在 18 小时内达到 25.4%的测试误差。对于语义分割任务,DSO-NAS 与在 PASCAL VOC 数据集上手动设计的架构相比也取得了有竞争力的结果。代码可在 https://github.com/XinbangZhang/DSO-NAS 上获得。

相似文献

1
You Only Search Once: Single Shot Neural Architecture Search via Direct Sparse Optimization.你只需搜索一次:通过直接稀疏优化的单镜头神经架构搜索。
IEEE Trans Pattern Anal Mach Intell. 2021 Sep;43(9):2891-2904. doi: 10.1109/TPAMI.2020.3020300. Epub 2021 Aug 4.
2
MIGO-NAS: Towards Fast and Generalizable Neural Architecture Search.MIGO-NAS:迈向快速且可泛化的神经架构搜索。
IEEE Trans Pattern Anal Mach Intell. 2021 Sep;43(9):2936-2952. doi: 10.1109/TPAMI.2021.3065138. Epub 2021 Aug 4.
3
One-Shot Neural Architecture Search: Maximising Diversity to Overcome Catastrophic Forgetting.单次神经架构搜索:通过最大化多样性克服灾难性遗忘。
IEEE Trans Pattern Anal Mach Intell. 2021 Sep;43(9):2921-2935. doi: 10.1109/TPAMI.2020.3035351. Epub 2021 Aug 4.
4
Sample-Efficient Neural Architecture Search by Learning Actions for Monte Carlo Tree Search.基于学习动作的蒙特卡洛树搜索的高效神经架构搜索。
IEEE Trans Pattern Anal Mach Intell. 2022 Sep;44(9):5503-5515. doi: 10.1109/TPAMI.2021.3071343. Epub 2022 Aug 4.
5
RelativeNAS: Relative Neural Architecture Search via Slow-Fast Learning.相对NAS:通过快慢学习进行相对神经架构搜索
IEEE Trans Neural Netw Learn Syst. 2023 Jan;34(1):475-489. doi: 10.1109/TNNLS.2021.3096658. Epub 2023 Jan 5.
6
Deeply Supervised Block-Wise Neural Architecture Search.深度监督的逐块神经架构搜索
IEEE Trans Neural Netw Learn Syst. 2025 Feb;36(2):2451-2464. doi: 10.1109/TNNLS.2023.3347542. Epub 2025 Feb 6.
7
One-Shot Neural Architecture Search by Dynamically Pruning Supernet in Hierarchical Order.分层动态剪枝超网的单步神经架构搜索。
Int J Neural Syst. 2021 Jul;31(7):2150029. doi: 10.1142/S0129065721500295. Epub 2021 Jun 14.
8
MNGNAS: Distilling Adaptive Combination of Multiple Searched Networks for One-Shot Neural Architecture Search.MNGNAS:用于一次性神经架构搜索的多个搜索网络的自适应组合提取
IEEE Trans Pattern Anal Mach Intell. 2023 Nov;45(11):13489-13508. doi: 10.1109/TPAMI.2023.3293885. Epub 2023 Oct 3.
9
FNA++: Fast Network Adaptation via Parameter Remapping and Architecture Search.FNA++:通过参数重映射和架构搜索实现快速网络自适应。
IEEE Trans Pattern Anal Mach Intell. 2021 Sep;43(9):2990-3004. doi: 10.1109/TPAMI.2020.3044416. Epub 2021 Aug 4.
10
Neural Architecture Transfer.神经架构迁移。
IEEE Trans Pattern Anal Mach Intell. 2021 Sep;43(9):2971-2989. doi: 10.1109/TPAMI.2021.3052758. Epub 2021 Aug 4.