Louati Hassen, Louati Ali, Bechikh Slim, Kariri Elham
SMART Lab, University of Tunis, ISG, Tunis, Tunisia.
Department of Information Systems, College of Computer Engineering and Sciences, Prince Sattam bin Abdulaziz University, 11942 Al-Kharj, Saudi Arabia.
J Supercomput. 2023 Apr 25:1-34. doi: 10.1007/s11227-023-05273-5.
Remarkable advancements have been achieved in machine learning and computer vision through the utilization of deep neural networks. Among the most advantageous of these networks is the convolutional neural network (CNN). It has been used in pattern recognition, medical diagnosis, and signal processing, among other things. Actually, for these networks, the challenge of choosing hyperparameters is of utmost importance. The reason behind this is that as the number of layers rises, the search space grows exponentially. In addition, all known classical and evolutionary pruning algorithms require a trained or built architecture as input. During the design phase, none of them consider the process of pruning. In order to assess the effectiveness and efficiency of any architecture created, pruning of channels must be carried out before transmitting the dataset and computing classification errors. For instance, following pruning, an architecture of medium quality in terms of classification may transform into an architecture that is both highly light and accurate, and vice versa. There exist countless potential scenarios that could occur, which prompted us to develop a bi-level optimization approach for the entire process. The upper level involves generating the architecture while the lower level optimizes channel pruning. Evolutionary algorithms (EAs) have proven effective in bi-level optimization, leading us to adopt the co-evolutionary migration-based algorithm as a search engine for our bi-level architectural optimization problem in this research. Our proposed method, CNN-D-P (bi-level CNN design and pruning), was tested on the widely used image classification benchmark datasets, CIFAR-10, CIFAR-100 and ImageNet. Our suggested technique is validated by means of a set of comparison tests with regard to relevant state-of-the-art architectures.
通过利用深度神经网络,机器学习和计算机视觉取得了显著进展。其中最具优势的网络之一是卷积神经网络(CNN)。它已被用于模式识别、医学诊断和信号处理等领域。实际上,对于这些网络而言,选择超参数的挑战至关重要。其背后的原因是,随着层数的增加,搜索空间呈指数增长。此外,所有已知的经典和进化剪枝算法都需要一个经过训练或构建好的架构作为输入。在设计阶段,它们都没有考虑剪枝过程。为了评估所创建的任何架构的有效性和效率,必须在传输数据集和计算分类误差之前进行通道剪枝。例如,经过剪枝后,在分类方面质量中等的架构可能会转变为既高度轻量化又准确的架构,反之亦然。可能会出现无数种潜在情况,这促使我们为整个过程开发一种双层优化方法。上层涉及生成架构,而下层则优化通道剪枝。进化算法(EAs)已在双层优化中证明有效,这使我们在本研究中采用基于协同进化迁移的算法作为我们双层架构优化问题的搜索引擎。我们提出的方法CNN-D-P(双层CNN设计与剪枝)在广泛使用的图像分类基准数据集CIFAR-10、CIFAR-100和ImageNet上进行了测试。我们建议的技术通过与相关的最先进架构进行一组比较测试来验证。