Zhang Xu, Liu Shuai, Wang Xueli, Li Yumei
School of Mathematics and Statistics, Beijing Technology and Business University, Beijing, China.
Sci Rep. 2024 Jan 27;14(1):2291. doi: 10.1038/s41598-024-52945-0.
In recent years, deep neural networks have evolved rapidly in engineering technology, with models becoming larger and deeper. However, for most companies, developing large models is extremely costly and highly risky. Researchers usually focus on the performance of the model, neglecting its cost and accessibility. In fact, most regular business scenarios do not require high-level AI. A simple and inexpensive modeling method for fulfilling certain demands for practical applications of AI is needed. In this paper, a Fragmented neural network method is proposed. Inspired by the random forest algorithm, both the samples and features are randomly sampled on image data. Images are randomly split into smaller pieces. Weak neural networks are trained using these fragmented images, and many weak neural networks are then ensembled to build a strong neural network by voting. In this way, sufficient accuracy is achieved while reducing the complexity and data volume of each base learner, enabling mass production through parallel and distributed computing. By conducting experiments on the MNIST and CIFAR10 datasets, we build a model pool using FNN, CNN, DenseNet, and ResNet as the basic network structure. We find that the accuracy of the ensemble weak network is significantly higher than that of each base learner. Meanwhile, the accuracy of the ensemble network is highly dependent on the performance of each base learner. The accuracy of the ensemble network is comparable to or even exceeds that of the full model and has better robustness. Unlike other similar studies, we do not pursue SOTA models. Instead, we achieved results close to the full model with a smaller number of parameters and amount of data.
近年来,深度神经网络在工程技术领域发展迅速,模型变得越来越大、越来越深。然而,对于大多数公司来说,开发大型模型成本极高且风险很大。研究人员通常专注于模型的性能,而忽略了其成本和可访问性。事实上,大多数常规业务场景并不需要高级人工智能。因此,需要一种简单且低成本的建模方法来满足人工智能实际应用的某些需求。本文提出了一种碎片化神经网络方法。受随机森林算法的启发,对图像数据的样本和特征都进行随机采样。图像被随机分割成更小的块。使用这些碎片化图像训练弱神经网络,然后通过投票将许多弱神经网络集成起来构建一个强神经网络。通过这种方式,在降低每个基础学习器的复杂度和数据量的同时,实现了足够的精度,从而能够通过并行和分布式计算进行大规模生产。通过在MNIST和CIFAR10数据集上进行实验,我们以FNN、CNN、DenseNet和ResNet作为基本网络结构构建了一个模型池。我们发现,集成弱网络的准确率显著高于每个基础学习器。同时,集成网络的准确率高度依赖于每个基础学习器的性能。集成网络的准确率与完整模型相当甚至超过完整模型,并且具有更好的鲁棒性。与其他类似研究不同,我们不追求最先进的模型。相反,我们用较少的参数数量和数据量取得了接近完整模型的结果。