• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

整数值神经网络的混合整数规划最优训练。

Optimal training of integer-valued neural networks with mixed integer programming.

机构信息

Algorithmics Group, Faculty of Electrical Engineering, Mathematics and Computer Science, Delft University of Technology, Delft, The Netherlands.

出版信息

PLoS One. 2023 Feb 1;18(2):e0261029. doi: 10.1371/journal.pone.0261029. eCollection 2023.

DOI:10.1371/journal.pone.0261029
PMID:36724189
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9891529/
Abstract

Recent work has shown potential in using Mixed Integer Programming (MIP) solvers to optimize certain aspects of neural networks (NNs). However the intriguing approach of training NNs with MIP solvers is under-explored. State-of-the-art-methods to train NNs are typically gradient-based and require significant data, computation on GPUs, and extensive hyper-parameter tuning. In contrast, training with MIP solvers does not require GPUs or heavy hyper-parameter tuning, but currently cannot handle anything but small amounts of data. This article builds on recent advances that train binarized NNs using MIP solvers. We go beyond current work by formulating new MIP models which improve training efficiency and which can train the important class of integer-valued neural networks (INNs). We provide two novel methods to further the potential significance of using MIP to train NNs. The first method optimizes the number of neurons in the NN while training. This reduces the need for deciding on network architecture before training. The second method addresses the amount of training data which MIP can feasibly handle: we provide a batch training method that dramatically increases the amount of data that MIP solvers can use to train. We thus provide a promising step towards using much more data than before when training NNs using MIP models. Experimental results on two real-world data-limited datasets demonstrate that our approach strongly outperforms the previous state of the art in training NN with MIP, in terms of accuracy, training time and amount of data. Our methodology is proficient at training NNs when minimal training data is available, and at training with minimal memory requirements-which is potentially valuable for deploying to low-memory devices.

摘要

最近的工作表明,使用混合整数规划(MIP)求解器来优化神经网络(NN)的某些方面具有潜力。然而,使用 MIP 求解器训练神经网络的这种有趣方法还没有得到充分的探索。训练神经网络的最新方法通常是基于梯度的,需要大量的数据、GPU 上的计算和广泛的超参数调整。相比之下,使用 MIP 求解器进行训练不需要 GPU 或大量的超参数调整,但目前只能处理少量的数据。本文基于使用 MIP 求解器训练二值神经网络的最新进展。我们通过制定新的 MIP 模型来超越当前的工作,这些模型可以提高训练效率,并可以训练重要的整数神经网络(INN)。我们提供了两种新的方法来进一步提高使用 MIP 训练神经网络的潜在意义。第一种方法是在训练过程中优化神经网络中的神经元数量。这减少了在训练前决定网络架构的需要。第二种方法解决了 MIP 可以处理的训练数据量的问题:我们提供了一种批量训练方法,可以显著增加 MIP 求解器可以用于训练的数据量。因此,我们朝着使用比以前更多的数据来训练使用 MIP 模型的神经网络迈出了有希望的一步。在两个真实的数据集上的实验结果表明,在使用 MIP 训练 NN 方面,我们的方法在准确性、训练时间和数据量方面都明显优于之前的最先进方法。当可用的训练数据很少时,我们的方法能够熟练地训练神经网络,并且需要最小的内存要求-这对于部署到低内存设备来说是潜在有价值的。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8bdb/9891529/db7f5ec49718/pone.0261029.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8bdb/9891529/2fba0088ad63/pone.0261029.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8bdb/9891529/db7f5ec49718/pone.0261029.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8bdb/9891529/2fba0088ad63/pone.0261029.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/8bdb/9891529/db7f5ec49718/pone.0261029.g002.jpg

相似文献

1
Optimal training of integer-valued neural networks with mixed integer programming.整数值神经网络的混合整数规划最优训练。
PLoS One. 2023 Feb 1;18(2):e0261029. doi: 10.1371/journal.pone.0261029. eCollection 2023.
2
High frequency accuracy and loss data of random neural networks trained on image datasets.在图像数据集上训练的随机神经网络的高频精度和损失数据。
Data Brief. 2022 Jan 5;40:107780. doi: 10.1016/j.dib.2021.107780. eCollection 2022 Feb.
3
A Gradient-Guided Evolutionary Approach to Training Deep Neural Networks.基于梯度的深度神经网络进化训练方法。
IEEE Trans Neural Netw Learn Syst. 2022 Sep;33(9):4861-4875. doi: 10.1109/TNNLS.2021.3061630. Epub 2022 Aug 31.
4
A constructive algorithm for training cooperative neural network ensembles.一种用于训练协作神经网络集成的构造性算法。
IEEE Trans Neural Netw. 2003;14(4):820-34. doi: 10.1109/TNN.2003.813832.
5
Should We Stop Developing Heuristics and Only Rely on Mixed Integer Programming Solvers in Automated Test Assembly? A Rejoinder to van der Linden and Li (2016).我们应该停止开发启发式方法,而仅在自动测试组卷中依赖混合整数规划求解器吗?对范德林登和李(2016年)的回应
Appl Psychol Meas. 2017 May;41(3):227-240. doi: 10.1177/0146621617695523. Epub 2017 Mar 14.
6
A Survey of Stochastic Computing Neural Networks for Machine Learning Applications.用于机器学习应用的随机计算神经网络调查。
IEEE Trans Neural Netw Learn Syst. 2021 Jul;32(7):2809-2824. doi: 10.1109/TNNLS.2020.3009047. Epub 2021 Jul 6.
7
Residual DNN: training diffractive deep neural networks via learnable light shortcuts.残留 DNN:通过可学习的光捷径训练衍射深度神经网络。
Opt Lett. 2020 May 15;45(10):2688-2691. doi: 10.1364/OL.389696.
8
Comment on Three-Element Item Selection Procedures for Multiple Forms Assembly: An Item Matching Approach.关于多份试卷组卷的三要素选题程序的评论:一种试题匹配方法
Appl Psychol Meas. 2016 Nov;40(8):641-649. doi: 10.1177/0146621616664075. Epub 2016 Aug 20.
9
Non-linear Memristive Synaptic Dynamics for Efficient Unsupervised Learning in Spiking Neural Networks.用于脉冲神经网络中高效无监督学习的非线性忆阻突触动力学
Front Neurosci. 2021 Feb 1;15:580909. doi: 10.3389/fnins.2021.580909. eCollection 2021.
10
Bagging and boosting negatively correlated neural networks.装袋法和提升法与神经网络呈负相关。
IEEE Trans Syst Man Cybern B Cybern. 2008 Jun;38(3):771-84. doi: 10.1109/TSMCB.2008.922055.